2026-03-06T13:46:09.197 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-06T13:46:09.203 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:46:09.226 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278 branch: cobaltcore-storage-v19.2.3-fasttrack-3 description: orch:cephadm:workunits/{0-distro/centos_9.stream_runc agent/on mon_election/connectivity task/test_extra_daemon_features} email: null first_in_suite: false flavor: default job_id: '278' last_in_suite: false machine_type: vps name: irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-3 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: global: mon election default strategy: 3 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: true mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 install: ceph: flavor: default sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64 selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-19.2.3-fasttrack-3-no-nvme-loop sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 seed: 6609 sha1: c24117fd5525679b799527bc1bd1f1dd0a2db5e2 sleep_before_teardown: 0 subset: 1/64 suite: orch:cephadm:workunits suite_branch: tt-19.2.3-fasttrack-3-no-nvme-loop suite_path: /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 5726a36c3452e5b72190cfceba828abc62c819b7 targets: vm00.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJSHBHjwgbTIcYCFpAlrOr5Q9IhC2C9iMbqM8xYcoYBNCleqFBbWge/+Ni0yTARqgxpd5JmdPt+7FMuLt8BTZrc= vm01.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBN8uOmejwP/RfIkqSRHdsbV7MAmRlpSu7Ml+srSW0yu80la0iOtSa4jCbOP5daiHverzKAuqPrDI89BZrk6tTlo= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - install: null - cephadm: null - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon - custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest - custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z - cephadm.wait_for_service: service: mon - cephadm.wait_for_service: service: container.foo - cephadm.wait_for_service: service: container.bar - exec: host.a: - 'set -ex FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) sleep 60 # check extra container and entrypoint args written to mon unit run file grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run # check that custom container properly wrote content to file. # This requires the custom config, extra container args, and # entrypoint args to all be working in order for this to have # been written. The container entrypoint was set up with custom_configs, # the content and where to write to with the entrypoint args, and the mounting # of the /etc/cephadm_testing dir with extra container args grep "testing_custom_containers" /etc/cephadm_testing/testing.txt # Verify that container bar''s init containers and primary container # ran successfully dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name ''container.bar.*'') test -n "$dir" grep ok ${dir}/data/primary.txt grep from=centos ${dir}/data/from.txt test -s ${dir}/data/presized.dat ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-06_13:20:18 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.43333 2026-03-06T13:46:09.226 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa; will attempt to use it 2026-03-06T13:46:09.226 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_5726a36c3452e5b72190cfceba828abc62c819b7/qa/tasks 2026-03-06T13:46:09.226 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-06T13:46:09.227 INFO:teuthology.task.internal:Saving configuration 2026-03-06T13:46:09.233 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-06T13:46:09.234 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-06T13:46:09.240 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm00.local', 'description': '/archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 12:44:55.483557', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:00', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJSHBHjwgbTIcYCFpAlrOr5Q9IhC2C9iMbqM8xYcoYBNCleqFBbWge/+Ni0yTARqgxpd5JmdPt+7FMuLt8BTZrc='} 2026-03-06T13:46:09.245 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm01.local', 'description': '/archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-06 12:44:55.483975', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:01', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBN8uOmejwP/RfIkqSRHdsbV7MAmRlpSu7Ml+srSW0yu80la0iOtSa4jCbOP5daiHverzKAuqPrDI89BZrk6tTlo='} 2026-03-06T13:46:09.245 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-06T13:46:09.245 INFO:teuthology.task.internal:roles: ubuntu@vm00.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0'] 2026-03-06T13:46:09.245 INFO:teuthology.task.internal:roles: ubuntu@vm01.local - ['host.b', 'mon.b', 'mgr.b', 'osd.1'] 2026-03-06T13:46:09.245 INFO:teuthology.run_tasks:Running task console_log... 2026-03-06T13:46:09.251 DEBUG:teuthology.task.console_log:vm00 does not support IPMI; excluding 2026-03-06T13:46:09.261 DEBUG:teuthology.task.console_log:vm01 does not support IPMI; excluding 2026-03-06T13:46:09.261 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f98a1bb3d90>, signals=[15]) 2026-03-06T13:46:09.261 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-06T13:46:09.262 INFO:teuthology.task.internal:Opening connections... 2026-03-06T13:46:09.262 DEBUG:teuthology.task.internal:connecting to ubuntu@vm00.local 2026-03-06T13:46:09.263 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:46:09.322 DEBUG:teuthology.task.internal:connecting to ubuntu@vm01.local 2026-03-06T13:46:09.322 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:46:09.381 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-06T13:46:09.382 DEBUG:teuthology.orchestra.run.vm00:> uname -m 2026-03-06T13:46:09.426 INFO:teuthology.orchestra.run.vm00.stdout:x86_64 2026-03-06T13:46:09.427 DEBUG:teuthology.orchestra.run.vm00:> cat /etc/os-release 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:NAME="CentOS Stream" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:VERSION="9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:ID="centos" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:VERSION_ID="9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:ANSI_COLOR="0;31" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:46:09.486 INFO:teuthology.orchestra.run.vm00.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:46:09.486 INFO:teuthology.lock.ops:Updating vm00.local on lock server 2026-03-06T13:46:09.491 DEBUG:teuthology.orchestra.run.vm01:> uname -m 2026-03-06T13:46:09.513 INFO:teuthology.orchestra.run.vm01.stdout:x86_64 2026-03-06T13:46:09.513 DEBUG:teuthology.orchestra.run.vm01:> cat /etc/os-release 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:NAME="CentOS Stream" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:VERSION="9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:ID="centos" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:ID_LIKE="rhel fedora" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:VERSION_ID="9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:PLATFORM_ID="platform:el9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:ANSI_COLOR="0;31" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:LOGO="fedora-logo-icon" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:HOME_URL="https://centos.org/" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-06T13:46:09.567 INFO:teuthology.orchestra.run.vm01.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-06T13:46:09.568 INFO:teuthology.lock.ops:Updating vm01.local on lock server 2026-03-06T13:46:09.572 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-06T13:46:09.573 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-06T13:46:09.574 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-06T13:46:09.574 DEBUG:teuthology.orchestra.run.vm00:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:46:09.576 DEBUG:teuthology.orchestra.run.vm01:> test '!' -e /home/ubuntu/cephtest 2026-03-06T13:46:09.621 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-06T13:46:09.621 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-06T13:46:09.622 DEBUG:teuthology.orchestra.run.vm00:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:46:09.634 DEBUG:teuthology.orchestra.run.vm01:> test -z $(ls -A /var/lib/ceph) 2026-03-06T13:46:09.651 INFO:teuthology.orchestra.run.vm00.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:46:09.678 INFO:teuthology.orchestra.run.vm01.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-06T13:46:09.679 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-06T13:46:09.686 DEBUG:teuthology.orchestra.run.vm00:> test -e /ceph-qa-ready 2026-03-06T13:46:09.706 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:09.892 DEBUG:teuthology.orchestra.run.vm01:> test -e /ceph-qa-ready 2026-03-06T13:46:09.906 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:10.092 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-06T13:46:10.094 INFO:teuthology.task.internal:Creating test directory... 2026-03-06T13:46:10.094 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:46:10.096 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-06T13:46:10.111 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-06T13:46:10.113 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-06T13:46:10.114 INFO:teuthology.task.internal:Creating archive directory... 2026-03-06T13:46:10.114 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:46:10.154 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-06T13:46:10.173 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-06T13:46:10.174 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-06T13:46:10.174 DEBUG:teuthology.orchestra.run.vm00:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:46:10.229 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:10.229 DEBUG:teuthology.orchestra.run.vm01:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-06T13:46:10.245 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:46:10.245 DEBUG:teuthology.orchestra.run.vm00:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:46:10.271 DEBUG:teuthology.orchestra.run.vm01:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-06T13:46:10.296 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:46:10.305 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:46:10.316 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:46:10.327 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-06T13:46:10.328 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-06T13:46:10.330 INFO:teuthology.task.internal:Configuring sudo... 2026-03-06T13:46:10.330 DEBUG:teuthology.orchestra.run.vm00:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:46:10.349 DEBUG:teuthology.orchestra.run.vm01:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-06T13:46:10.396 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-06T13:46:10.398 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-06T13:46:10.399 DEBUG:teuthology.orchestra.run.vm00:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:46:10.413 DEBUG:teuthology.orchestra.run.vm01:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-06T13:46:10.453 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:46:10.492 DEBUG:teuthology.orchestra.run.vm00:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:46:10.549 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:46:10.549 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:46:10.612 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:46:10.639 DEBUG:teuthology.orchestra.run.vm01:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:46:10.697 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:46:10.697 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-06T13:46:10.756 DEBUG:teuthology.orchestra.run.vm00:> sudo service rsyslog restart 2026-03-06T13:46:10.758 DEBUG:teuthology.orchestra.run.vm01:> sudo service rsyslog restart 2026-03-06T13:46:10.783 INFO:teuthology.orchestra.run.vm00.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:46:10.827 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:46:11.066 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-06T13:46:11.067 INFO:teuthology.task.internal:Starting timer... 2026-03-06T13:46:11.067 INFO:teuthology.run_tasks:Running task pcp... 2026-03-06T13:46:11.070 INFO:teuthology.run_tasks:Running task selinux... 2026-03-06T13:46:11.072 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-06T13:46:11.072 INFO:teuthology.task.selinux:Excluding vm00: VMs are not yet supported 2026-03-06T13:46:11.072 INFO:teuthology.task.selinux:Excluding vm01: VMs are not yet supported 2026-03-06T13:46:11.072 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-06T13:46:11.072 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-06T13:46:11.072 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-06T13:46:11.072 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-06T13:46:11.073 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-06T13:46:11.074 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-06T13:46:11.075 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-06T13:46:11.561 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-06T13:46:11.567 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-06T13:46:11.567 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventory75pz6k7o --limit vm00.local,vm01.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-06T13:48:18.338 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm00.local'), Remote(name='ubuntu@vm01.local')] 2026-03-06T13:48:18.338 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm00.local' 2026-03-06T13:48:18.338 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm00.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:48:18.405 DEBUG:teuthology.orchestra.run.vm00:> true 2026-03-06T13:48:18.488 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm00.local' 2026-03-06T13:48:18.489 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm01.local' 2026-03-06T13:48:18.489 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm01.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-06T13:48:18.552 DEBUG:teuthology.orchestra.run.vm01:> true 2026-03-06T13:48:18.634 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm01.local' 2026-03-06T13:48:18.634 INFO:teuthology.run_tasks:Running task clock... 2026-03-06T13:48:18.637 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-06T13:48:18.637 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:48:18.637 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:48:18.639 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-06T13:48:18.639 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:48:18.673 INFO:teuthology.orchestra.run.vm00.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:48:18.690 INFO:teuthology.orchestra.run.vm00.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:48:18.708 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-06T13:48:18.724 INFO:teuthology.orchestra.run.vm00.stderr:sudo: ntpd: command not found 2026-03-06T13:48:18.724 INFO:teuthology.orchestra.run.vm01.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-06T13:48:18.736 INFO:teuthology.orchestra.run.vm00.stdout:506 Cannot talk to daemon 2026-03-06T13:48:18.752 INFO:teuthology.orchestra.run.vm00.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:48:18.752 INFO:teuthology.orchestra.run.vm01.stderr:sudo: ntpd: command not found 2026-03-06T13:48:18.764 INFO:teuthology.orchestra.run.vm01.stdout:506 Cannot talk to daemon 2026-03-06T13:48:18.768 INFO:teuthology.orchestra.run.vm00.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:48:18.779 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-06T13:48:18.797 INFO:teuthology.orchestra.run.vm01.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-06T13:48:18.819 INFO:teuthology.orchestra.run.vm00.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:48:18.820 INFO:teuthology.orchestra.run.vm00.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:48:18.820 INFO:teuthology.orchestra.run.vm00.stdout:=============================================================================== 2026-03-06T13:48:18.838 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:48:18.841 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:48:18.841 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T13:48:18.841 INFO:teuthology.run_tasks:Running task pexec... 2026-03-06T13:48:18.843 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-06T13:48:18.843 DEBUG:teuthology.orchestra.run.vm00:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:48:18.844 DEBUG:teuthology.orchestra.run.vm01:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-06T13:48:18.863 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo dnf remove nvme-cli -y 2026-03-06T13:48:18.863 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:48:18.863 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.863 DEBUG:teuthology.task.pexec:ubuntu@vm00.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.863 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm00.local 2026-03-06T13:48:18.863 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:48:18.863 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:48:18.863 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.863 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.884 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf remove nvme-cli -y 2026-03-06T13:48:18.885 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:48:18.885 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.885 DEBUG:teuthology.task.pexec:ubuntu@vm01.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.885 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm01.local 2026-03-06T13:48:18.885 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-06T13:48:18.885 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-06T13:48:18.885 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:18.885 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-06T13:48:19.061 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: nvme-cli 2026-03-06T13:48:19.061 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:48:19.064 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:48:19.064 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:48:19.064 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:48:19.085 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: nvme-cli 2026-03-06T13:48:19.085 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:48:19.087 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:48:19.088 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:48:19.088 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:48:19.409 INFO:teuthology.orchestra.run.vm01.stdout:Last metadata expiration check: 0:01:18 ago on Fri 06 Mar 2026 01:47:01 PM CET. 2026-03-06T13:48:19.498 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Install 7 Packages 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 6.3 M 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Installed size: 24 M 2026-03-06T13:48:19.499 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-06T13:48:19.528 INFO:teuthology.orchestra.run.vm00.stdout:Last metadata expiration check: 0:01:25 ago on Fri 06 Mar 2026 01:46:54 PM CET. 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:Installing: 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:Installing dependencies: 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:48:19.643 INFO:teuthology.orchestra.run.vm00.stdout:Install 7 Packages 2026-03-06T13:48:19.644 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:19.644 INFO:teuthology.orchestra.run.vm00.stdout:Total download size: 6.3 M 2026-03-06T13:48:19.644 INFO:teuthology.orchestra.run.vm00.stdout:Installed size: 24 M 2026-03-06T13:48:19.644 INFO:teuthology.orchestra.run.vm00.stdout:Downloading Packages: 2026-03-06T13:48:20.730 INFO:teuthology.orchestra.run.vm01.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 264 kB/s | 44 kB 00:00 2026-03-06T13:48:20.776 INFO:teuthology.orchestra.run.vm01.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 338 kB/s | 72 kB 00:00 2026-03-06T13:48:20.795 INFO:teuthology.orchestra.run.vm00.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 87 kB/s | 44 kB 00:00 2026-03-06T13:48:20.825 INFO:teuthology.orchestra.run.vm01.stdout:(3/7): nvme-cli-2.16-1.el9.x86_64.rpm 4.4 MB/s | 1.2 MB 00:00 2026-03-06T13:48:20.831 INFO:teuthology.orchestra.run.vm00.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 133 kB/s | 72 kB 00:00 2026-03-06T13:48:20.885 INFO:teuthology.orchestra.run.vm01.stdout:(4/7): python3-kmod-0.9-32.el9.x86_64.rpm 541 kB/s | 84 kB 00:00 2026-03-06T13:48:20.960 INFO:teuthology.orchestra.run.vm00.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 509 kB/s | 84 kB 00:00 2026-03-06T13:48:20.964 INFO:teuthology.orchestra.run.vm01.stdout:(5/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 803 kB/s | 150 kB 00:00 2026-03-06T13:48:21.000 INFO:teuthology.orchestra.run.vm00.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 889 kB/s | 150 kB 00:00 2026-03-06T13:48:21.056 INFO:teuthology.orchestra.run.vm01.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 3.6 MB/s | 837 kB 00:00 2026-03-06T13:48:21.071 INFO:teuthology.orchestra.run.vm00.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 1.5 MB/s | 1.2 MB 00:00 2026-03-06T13:48:21.147 INFO:teuthology.orchestra.run.vm01.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 15 MB/s | 4.0 MB 00:00 2026-03-06T13:48:21.147 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:48:21.147 INFO:teuthology.orchestra.run.vm01.stdout:Total 3.8 MB/s | 6.3 MB 00:01 2026-03-06T13:48:21.215 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:48:21.223 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:48:21.223 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:48:21.284 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:48:21.284 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:48:21.285 INFO:teuthology.orchestra.run.vm00.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 2.5 MB/s | 837 kB 00:00 2026-03-06T13:48:21.436 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:48:21.446 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T13:48:21.457 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T13:48:21.464 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:48:21.472 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:48:21.473 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:48:21.528 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:48:21.662 INFO:teuthology.orchestra.run.vm01.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T13:48:21.667 INFO:teuthology.orchestra.run.vm01.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:48:21.760 INFO:teuthology.orchestra.run.vm00.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 5.2 MB/s | 4.0 MB 00:00 2026-03-06T13:48:21.760 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:48:21.760 INFO:teuthology.orchestra.run.vm00.stdout:Total 3.0 MB/s | 6.3 MB 00:02 2026-03-06T13:48:21.861 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:48:21.870 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:48:21.870 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:48:21.953 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:48:21.953 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:48:22.039 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:48:22.039 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:48:22.039 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:22.150 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:48:22.161 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-06T13:48:22.177 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-06T13:48:22.187 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:48:22.197 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:48:22.199 INFO:teuthology.orchestra.run.vm00.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:48:22.260 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-06T13:48:22.419 INFO:teuthology.orchestra.run.vm00.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-06T13:48:22.424 INFO:teuthology.orchestra.run.vm00.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T13:48:22.615 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:22.720 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:48:22.828 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-06T13:48:22.828 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-06T13:48:22.828 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:22.842 DEBUG:teuthology.parallel:result is None 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-06T13:48:23.439 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout:Installed: 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:23.531 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:48:23.625 DEBUG:teuthology.parallel:result is None 2026-03-06T13:48:23.625 INFO:teuthology.run_tasks:Running task install... 2026-03-06T13:48:23.627 DEBUG:teuthology.task.install:project ceph 2026-03-06T13:48:23.627 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:48:23.627 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}} 2026-03-06T13:48:23.627 INFO:teuthology.task.install:Using flavor: default 2026-03-06T13:48:23.629 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-06T13:48:23.629 INFO:teuthology.task.install:extra packages: [] 2026-03-06T13:48:23.629 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:48:23.629 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:48:23.629 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:48:23.629 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:48:23.630 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}]} 2026-03-06T13:48:23.630 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-47-gc24117fd552/el9.clyso/x86_64'}] 2026-03-06T13:48:23.630 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:48:23.630 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-03-06T13:48:23.665 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:48:23.665 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:48:23.665 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:48:23.665 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:48:23.735 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:48:23.735 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:48:23.740 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:48:23.741 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-06T13:48:23.803 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:48:23.803 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-06T13:48:23.815 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, s3cmd on remote rpm x86_64 2026-03-06T13:48:23.815 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean all 2026-03-06T13:48:23.979 INFO:teuthology.orchestra.run.vm01.stdout:41 files removed 2026-03-06T13:48:24.009 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:48:24.043 INFO:teuthology.orchestra.run.vm00.stdout:41 files removed 2026-03-06T13:48:24.073 DEBUG:teuthology.orchestra.run.vm00:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict s3cmd 2026-03-06T13:48:24.477 INFO:teuthology.orchestra.run.vm01.stdout:ceph 284 kB/s | 86 kB 00:00 2026-03-06T13:48:24.482 INFO:teuthology.orchestra.run.vm00.stdout:ceph 395 kB/s | 86 kB 00:00 2026-03-06T13:48:24.694 INFO:teuthology.orchestra.run.vm01.stdout:ceph-noarch 62 kB/s | 12 kB 00:00 2026-03-06T13:48:24.713 INFO:teuthology.orchestra.run.vm00.stdout:ceph-noarch 58 kB/s | 12 kB 00:00 2026-03-06T13:48:24.889 INFO:teuthology.orchestra.run.vm01.stdout:ceph-source 12 kB/s | 2.2 kB 00:00 2026-03-06T13:48:24.930 INFO:teuthology.orchestra.run.vm00.stdout:ceph-source 11 kB/s | 2.2 kB 00:00 2026-03-06T13:48:25.738 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - BaseOS 11 MB/s | 8.9 MB 00:00 2026-03-06T13:48:27.542 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - AppStream 23 MB/s | 27 MB 00:01 2026-03-06T13:48:29.117 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - BaseOS 2.1 MB/s | 8.9 MB 00:04 2026-03-06T13:48:30.959 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - AppStream 22 MB/s | 27 MB 00:01 2026-03-06T13:48:34.102 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - CRB 14 MB/s | 8.0 MB 00:00 2026-03-06T13:48:35.725 INFO:teuthology.orchestra.run.vm00.stdout:CentOS Stream 9 - Extras packages 25 kB/s | 20 kB 00:00 2026-03-06T13:48:36.165 INFO:teuthology.orchestra.run.vm00.stdout:Extra Packages for Enterprise Linux 55 MB/s | 20 MB 00:00 2026-03-06T13:48:38.589 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - CRB 974 kB/s | 8.0 MB 00:08 2026-03-06T13:48:39.693 INFO:teuthology.orchestra.run.vm01.stdout:CentOS Stream 9 - Extras packages 69 kB/s | 20 kB 00:00 2026-03-06T13:48:40.448 INFO:teuthology.orchestra.run.vm00.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-06T13:48:40.471 INFO:teuthology.orchestra.run.vm01.stdout:Extra Packages for Enterprise Linux 29 MB/s | 20 MB 00:00 2026-03-06T13:48:41.759 INFO:teuthology.orchestra.run.vm00.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:48:41.760 INFO:teuthology.orchestra.run.vm00.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:48:41.790 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout:Installing: 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:48:41.795 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout:Upgrading: 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout:Installing dependencies: 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:48:41.796 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout:Installing weak dependencies: 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:48:41.797 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:============================================================================================== 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:Install 148 Packages 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:Upgrade 2 Packages 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:Total download size: 214 M 2026-03-06T13:48:41.798 INFO:teuthology.orchestra.run.vm00.stdout:Downloading Packages: 2026-03-06T13:48:43.084 INFO:teuthology.orchestra.run.vm00.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 79 kB/s | 6.5 kB 00:00 2026-03-06T13:48:43.367 INFO:teuthology.orchestra.run.vm00.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 4.1 MB/s | 1.1 MB 00:00 2026-03-06T13:48:43.406 INFO:teuthology.orchestra.run.vm00.stdout:(3/150): ceph-immutable-object-cache-19.2.3-47. 3.6 MB/s | 145 kB 00:00 2026-03-06T13:48:43.559 INFO:teuthology.orchestra.run.vm00.stdout:(4/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 16 MB/s | 2.4 MB 00:00 2026-03-06T13:48:43.606 INFO:teuthology.orchestra.run.vm00.stdout:(5/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 9.1 MB/s | 5.5 MB 00:00 2026-03-06T13:48:43.637 INFO:teuthology.orchestra.run.vm00.stdout:(6/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 14 MB/s | 1.1 MB 00:00 2026-03-06T13:48:43.870 INFO:teuthology.orchestra.run.vm00.stdout:(7/150): ceph-common-19.2.3-47.gc24117fd552.el9 25 MB/s | 22 MB 00:00 2026-03-06T13:48:43.952 INFO:teuthology.orchestra.run.vm00.stdout:(8/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 14 MB/s | 4.7 MB 00:00 2026-03-06T13:48:43.995 INFO:teuthology.orchestra.run.vm00.stdout:(9/150): ceph-selinux-19.2.3-47.gc24117fd552.el 585 kB/s | 25 kB 00:00 2026-03-06T13:48:44.204 INFO:teuthology.orchestra.run.vm00.stdout:(10/150): ceph-osd-19.2.3-47.gc24117fd552.el9.c 30 MB/s | 17 MB 00:00 2026-03-06T13:48:44.258 INFO:teuthology.orchestra.run.vm00.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 630 kB/s | 34 kB 00:00 2026-03-06T13:48:44.337 INFO:teuthology.orchestra.run.vm00.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 12 MB/s | 998 kB 00:00 2026-03-06T13:48:44.387 INFO:teuthology.orchestra.run.vm00.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 3.3 MB/s | 163 kB 00:00 2026-03-06T13:48:44.441 INFO:teuthology.orchestra.run.vm00.stdout:(14/150): librados-devel-19.2.3-47.gc24117fd552 2.3 MB/s | 127 kB 00:00 2026-03-06T13:48:44.494 INFO:teuthology.orchestra.run.vm00.stdout:(15/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 17 MB/s | 11 MB 00:00 2026-03-06T13:48:44.516 INFO:teuthology.orchestra.run.vm00.stdout:(16/150): libradosstriper1-19.2.3-47.gc24117fd5 6.6 MB/s | 503 kB 00:00 2026-03-06T13:48:44.563 INFO:teuthology.orchestra.run.vm00.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 969 kB/s | 45 kB 00:00 2026-03-06T13:48:44.620 INFO:teuthology.orchestra.run.vm00.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 2.5 MB/s | 142 kB 00:00 2026-03-06T13:48:44.674 INFO:teuthology.orchestra.run.vm00.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 3.0 MB/s | 165 kB 00:00 2026-03-06T13:48:44.736 INFO:teuthology.orchestra.run.vm00.stdout:(20/150): python3-rados-19.2.3-47.gc24117fd552. 5.1 MB/s | 322 kB 00:00 2026-03-06T13:48:44.752 INFO:teuthology.orchestra.run.vm00.stdout:(21/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 21 MB/s | 5.4 MB 00:00 2026-03-06T13:48:44.773 INFO:teuthology.orchestra.run.vm00.stdout:(22/150): python3-rbd-19.2.3-47.gc24117fd552.el 8.1 MB/s | 303 kB 00:00 2026-03-06T13:48:44.785 INFO:teuthology.orchestra.run.vm00.stdout:(23/150): python3-rgw-19.2.3-47.gc24117fd552.el 2.9 MB/s | 100 kB 00:00 2026-03-06T13:48:44.810 INFO:teuthology.orchestra.run.vm00.stdout:(24/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 2.2 MB/s | 85 kB 00:00 2026-03-06T13:48:44.848 INFO:teuthology.orchestra.run.vm00.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 4.4 MB/s | 171 kB 00:00 2026-03-06T13:48:44.880 INFO:teuthology.orchestra.run.vm00.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 975 kB/s | 31 kB 00:00 2026-03-06T13:48:44.925 INFO:teuthology.orchestra.run.vm00.stdout:(27/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 3.3 MB/s | 150 kB 00:00 2026-03-06T13:48:44.931 INFO:teuthology.orchestra.run.vm01.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-03-06T13:48:44.983 INFO:teuthology.orchestra.run.vm00.stdout:(28/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 16 MB/s | 3.1 MB 00:00 2026-03-06T13:48:45.098 INFO:teuthology.orchestra.run.vm00.stdout:(29/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 22 MB/s | 3.8 MB 00:00 2026-03-06T13:48:45.127 INFO:teuthology.orchestra.run.vm00.stdout:(30/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 749 kB/s | 22 kB 00:00 2026-03-06T13:48:45.253 INFO:teuthology.orchestra.run.vm00.stdout:(31/150): ceph-test-19.2.3-47.gc24117fd552.el9. 39 MB/s | 50 MB 00:01 2026-03-06T13:48:45.255 INFO:teuthology.orchestra.run.vm00.stdout:(32/150): ceph-mgr-modules-core-19.2.3-47.gc241 1.9 MB/s | 252 kB 00:00 2026-03-06T13:48:45.285 INFO:teuthology.orchestra.run.vm00.stdout:(33/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.5 MB/s | 49 kB 00:00 2026-03-06T13:48:45.290 INFO:teuthology.orchestra.run.vm00.stdout:(34/150): ceph-prometheus-alerts-19.2.3-47.gc24 482 kB/s | 16 kB 00:00 2026-03-06T13:48:45.348 INFO:teuthology.orchestra.run.vm00.stdout:(35/150): ceph-mgr-diskprediction-local-19.2.3- 20 MB/s | 7.4 MB 00:00 2026-03-06T13:48:45.349 INFO:teuthology.orchestra.run.vm00.stdout:(36/150): ceph-volume-19.2.3-47.gc24117fd552.el 4.5 MB/s | 299 kB 00:00 2026-03-06T13:48:45.360 INFO:teuthology.orchestra.run.vm00.stdout:(37/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 11 MB/s | 769 kB 00:00 2026-03-06T13:48:45.546 INFO:teuthology.orchestra.run.vm00.stdout:(38/150): bzip2-1.0.8-11.el9.x86_64.rpm 276 kB/s | 55 kB 00:00 2026-03-06T13:48:45.561 INFO:teuthology.orchestra.run.vm00.stdout:(39/150): fuse-2.9.9-17.el9.x86_64.rpm 397 kB/s | 80 kB 00:00 2026-03-06T13:48:45.607 INFO:teuthology.orchestra.run.vm00.stdout:(40/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 670 kB/s | 40 kB 00:00 2026-03-06T13:48:45.626 INFO:teuthology.orchestra.run.vm00.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 1.1 MB/s | 72 kB 00:00 2026-03-06T13:48:45.646 INFO:teuthology.orchestra.run.vm00.stdout:(42/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.2 MB/s | 351 kB 00:00 2026-03-06T13:48:45.694 INFO:teuthology.orchestra.run.vm00.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 700 kB/s | 33 kB 00:00 2026-03-06T13:48:45.726 INFO:teuthology.orchestra.run.vm00.stdout:(44/150): libquadmath-11.5.0-14.el9.x86_64.rpm 1.8 MB/s | 184 kB 00:00 2026-03-06T13:48:45.751 INFO:teuthology.orchestra.run.vm00.stdout:(45/150): pciutils-3.7.0-7.el9.x86_64.rpm 1.6 MB/s | 93 kB 00:00 2026-03-06T13:48:45.782 INFO:teuthology.orchestra.run.vm00.stdout:(46/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 4.5 MB/s | 253 kB 00:00 2026-03-06T13:48:45.837 INFO:teuthology.orchestra.run.vm00.stdout:(47/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.4 MB/s | 794 kB 00:00 2026-03-06T13:48:45.843 INFO:teuthology.orchestra.run.vm00.stdout:(48/150): python3-ply-3.11-14.el9.noarch.rpm 1.7 MB/s | 106 kB 00:00 2026-03-06T13:48:45.890 INFO:teuthology.orchestra.run.vm00.stdout:(49/150): python3-requests-2.25.1-10.el9.noarch 2.6 MB/s | 126 kB 00:00 2026-03-06T13:48:45.898 INFO:teuthology.orchestra.run.vm00.stdout:(50/150): python3-cryptography-36.0.1-5.el9.x86 8.5 MB/s | 1.2 MB 00:00 2026-03-06T13:48:45.926 INFO:teuthology.orchestra.run.vm00.stdout:(51/150): python3-pycparser-2.20-6.el9.noarch.r 1.5 MB/s | 135 kB 00:00 2026-03-06T13:48:45.937 INFO:teuthology.orchestra.run.vm00.stdout:(52/150): python3-urllib3-1.26.5-7.el9.noarch.r 4.5 MB/s | 218 kB 00:00 2026-03-06T13:48:45.999 INFO:teuthology.orchestra.run.vm00.stdout:(53/150): smartmontools-7.2-10.el9.x86_64.rpm 5.4 MB/s | 556 kB 00:00 2026-03-06T13:48:46.001 INFO:teuthology.orchestra.run.vm00.stdout:(54/150): unzip-6.0-59.el9.x86_64.rpm 2.4 MB/s | 182 kB 00:00 2026-03-06T13:48:46.015 INFO:teuthology.orchestra.run.vm00.stdout:(55/150): zip-3.0-35.el9.x86_64.rpm 3.4 MB/s | 266 kB 00:00 2026-03-06T13:48:46.102 INFO:teuthology.orchestra.run.vm00.stdout:(56/150): flexiblas-3.0.4-9.el9.x86_64.rpm 293 kB/s | 30 kB 00:00 2026-03-06T13:48:46.149 INFO:teuthology.orchestra.run.vm00.stdout:(57/150): flexiblas-openblas-openmp-3.0.4-9.el9 320 kB/s | 15 kB 00:00 2026-03-06T13:48:46.184 INFO:teuthology.orchestra.run.vm00.stdout:(58/150): boost-program-options-1.75.0-13.el9.x 565 kB/s | 104 kB 00:00 2026-03-06T13:48:46.273 INFO:teuthology.orchestra.run.vm01.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:48:46.273 INFO:teuthology.orchestra.run.vm01.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-06T13:48:46.303 INFO:teuthology.orchestra.run.vm00.stdout:(59/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.3 MB/s | 160 kB 00:00 2026-03-06T13:48:46.304 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout:Installing: 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: bzip2 x86_64 1.0.8-11.el9 baseos 55 k 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 6.5 k 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.5 M 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 145 k 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 1.1 M 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 150 k 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 3.8 M 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 7.4 M 2026-03-06T13:48:46.308 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 49 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 11 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 50 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 299 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 769 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 34 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 998 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 127 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 295 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 165 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 322 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 303 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 100 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 85 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.1 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 171 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout:Upgrading: 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.4 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 3.2 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout:Installing dependencies: 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 22 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 31 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 2.4 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 252 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 4.7 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 17 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 16 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 25 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: fuse x86_64 2.9.9-17.el9 baseos 80 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 163 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-06T13:48:46.309 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 503 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 5.4 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: perl-Benchmark noarch 1.23-483.el9 appstream 26 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 45 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso ceph 142 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-06T13:48:46.310 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 189 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 90 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Installing weak dependencies: 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso ceph-noarch 22 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools x86_64 1:7.2-10.el9 baseos 556 k 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:============================================================================================== 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Install 148 Packages 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Upgrade 2 Packages 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm00.stdout:(60/150): libnbd-1.20.3-4.el9.x86_64.rpm 1.0 MB/s | 164 kB 00:00 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Total download size: 214 M 2026-03-06T13:48:46.311 INFO:teuthology.orchestra.run.vm01.stdout:Downloading Packages: 2026-03-06T13:48:46.361 INFO:teuthology.orchestra.run.vm00.stdout:(61/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 783 kB/s | 45 kB 00:00 2026-03-06T13:48:46.480 INFO:teuthology.orchestra.run.vm00.stdout:(62/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 2.0 MB/s | 246 kB 00:00 2026-03-06T13:48:46.547 INFO:teuthology.orchestra.run.vm00.stdout:(63/150): libxslt-1.1.34-12.el9.x86_64.rpm 3.4 MB/s | 233 kB 00:00 2026-03-06T13:48:46.553 INFO:teuthology.orchestra.run.vm00.stdout:(64/150): librdkafka-1.6.1-102.el9.x86_64.rpm 2.7 MB/s | 662 kB 00:00 2026-03-06T13:48:46.570 INFO:teuthology.orchestra.run.vm00.stdout:(65/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 5.4 MB/s | 3.0 MB 00:00 2026-03-06T13:48:46.625 INFO:teuthology.orchestra.run.vm00.stdout:(66/150): openblas-0.3.29-1.el9.x86_64.rpm 775 kB/s | 42 kB 00:00 2026-03-06T13:48:46.642 INFO:teuthology.orchestra.run.vm00.stdout:(67/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.0 MB/s | 292 kB 00:00 2026-03-06T13:48:46.643 INFO:teuthology.orchestra.run.vm00.stdout:(68/150): lua-5.4.4-4.el9.x86_64.rpm 2.1 MB/s | 188 kB 00:00 2026-03-06T13:48:46.691 INFO:teuthology.orchestra.run.vm00.stdout:(69/150): perl-Benchmark-1.23-483.el9.noarch.rp 540 kB/s | 26 kB 00:00 2026-03-06T13:48:46.760 INFO:teuthology.orchestra.run.vm00.stdout:(70/150): perl-Test-Harness-3.42-461.el9.noarch 2.5 MB/s | 295 kB 00:00 2026-03-06T13:48:46.886 INFO:teuthology.orchestra.run.vm00.stdout:(71/150): protobuf-3.14.0-17.el9.x86_64.rpm 5.2 MB/s | 1.0 MB 00:00 2026-03-06T13:48:46.967 INFO:teuthology.orchestra.run.vm00.stdout:(72/150): python3-devel-3.9.25-3.el9.x86_64.rpm 2.9 MB/s | 244 kB 00:00 2026-03-06T13:48:46.981 INFO:teuthology.orchestra.run.vm00.stdout:(73/150): openblas-openmp-0.3.29-1.el9.x86_64.r 15 MB/s | 5.3 MB 00:00 2026-03-06T13:48:47.051 INFO:teuthology.orchestra.run.vm00.stdout:(74/150): python3-jinja2-2.11.3-8.el9.noarch.rp 2.9 MB/s | 249 kB 00:00 2026-03-06T13:48:47.055 INFO:teuthology.orchestra.run.vm00.stdout:(75/150): python3-libstoragemgmt-1.10.1-1.el9.x 2.3 MB/s | 177 kB 00:00 2026-03-06T13:48:47.145 INFO:teuthology.orchestra.run.vm00.stdout:(76/150): python3-mako-1.1.4-6.el9.noarch.rpm 1.9 MB/s | 172 kB 00:00 2026-03-06T13:48:47.206 INFO:teuthology.orchestra.run.vm00.stdout:(77/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 7.9 MB/s | 1.2 MB 00:00 2026-03-06T13:48:47.207 INFO:teuthology.orchestra.run.vm00.stdout:(78/150): python3-markupsafe-1.1.1-12.el9.x86_6 558 kB/s | 35 kB 00:00 2026-03-06T13:48:47.264 INFO:teuthology.orchestra.run.vm00.stdout:(79/150): python3-babel-2.9.1-2.el9.noarch.rpm 12 MB/s | 6.0 MB 00:00 2026-03-06T13:48:47.331 INFO:teuthology.orchestra.run.vm00.stdout:(80/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 3.5 MB/s | 442 kB 00:00 2026-03-06T13:48:47.348 INFO:teuthology.orchestra.run.vm00.stdout:(81/150): python3-packaging-20.9-5.el9.noarch.r 927 kB/s | 77 kB 00:00 2026-03-06T13:48:47.403 INFO:teuthology.orchestra.run.vm00.stdout:(82/150): python3-protobuf-3.14.0-17.el9.noarch 3.6 MB/s | 267 kB 00:00 2026-03-06T13:48:47.416 INFO:teuthology.orchestra.run.vm00.stdout:(83/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 2.2 MB/s | 157 kB 00:00 2026-03-06T13:48:47.479 INFO:teuthology.orchestra.run.vm00.stdout:(84/150): python3-pyasn1-modules-0.4.8-7.el9.no 3.6 MB/s | 277 kB 00:00 2026-03-06T13:48:47.481 INFO:teuthology.orchestra.run.vm00.stdout:(85/150): python3-requests-oauthlib-1.3.0-12.el 835 kB/s | 54 kB 00:00 2026-03-06T13:48:47.546 INFO:teuthology.orchestra.run.vm00.stdout:(86/150): python3-toml-0.10.2-6.el9.noarch.rpm 646 kB/s | 42 kB 00:00 2026-03-06T13:48:47.587 INFO:teuthology.orchestra.run.vm00.stdout:(87/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 16 MB/s | 6.1 MB 00:00 2026-03-06T13:48:47.640 INFO:teuthology.orchestra.run.vm01.stdout:(1/150): ceph-19.2.3-47.gc24117fd552.el9.clyso. 74 kB/s | 6.5 kB 00:00 2026-03-06T13:48:47.647 INFO:teuthology.orchestra.run.vm00.stdout:(88/150): qatlib-25.08.0-2.el9.x86_64.rpm 2.3 MB/s | 240 kB 00:00 2026-03-06T13:48:47.648 INFO:teuthology.orchestra.run.vm00.stdout:(89/150): qatlib-service-25.08.0-2.el9.x86_64.r 606 kB/s | 37 kB 00:00 2026-03-06T13:48:47.714 INFO:teuthology.orchestra.run.vm00.stdout:(90/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.0 MB/s | 66 kB 00:00 2026-03-06T13:48:47.746 INFO:teuthology.orchestra.run.vm00.stdout:(91/150): socat-1.7.4.1-8.el9.x86_64.rpm 3.0 MB/s | 303 kB 00:00 2026-03-06T13:48:47.805 INFO:teuthology.orchestra.run.vm00.stdout:(92/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 2.0 MB/s | 189 kB 00:00 2026-03-06T13:48:47.823 INFO:teuthology.orchestra.run.vm00.stdout:(93/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 1.2 MB/s | 90 kB 00:00 2026-03-06T13:48:47.868 INFO:teuthology.orchestra.run.vm00.stdout:(94/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.0 MB/s | 64 kB 00:00 2026-03-06T13:48:47.909 INFO:teuthology.orchestra.run.vm01.stdout:(2/150): ceph-fuse-19.2.3-47.gc24117fd552.el9.c 4.3 MB/s | 1.1 MB 00:00 2026-03-06T13:48:47.911 INFO:teuthology.orchestra.run.vm00.stdout:(95/150): lua-devel-5.4.4-4.el9.x86_64.rpm 251 kB/s | 22 kB 00:00 2026-03-06T13:48:47.925 INFO:teuthology.orchestra.run.vm00.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 42 MB/s | 551 kB 00:00 2026-03-06T13:48:47.931 INFO:teuthology.orchestra.run.vm00.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 51 MB/s | 308 kB 00:00 2026-03-06T13:48:47.933 INFO:teuthology.orchestra.run.vm00.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 7.9 MB/s | 19 kB 00:00 2026-03-06T13:48:47.998 INFO:teuthology.orchestra.run.vm00.stdout:(99/150): protobuf-compiler-3.14.0-17.el9.x86_6 6.5 MB/s | 862 kB 00:00 2026-03-06T13:48:48.015 INFO:teuthology.orchestra.run.vm00.stdout:(100/150): libarrow-9.0.0-15.el9.x86_64.rpm 54 MB/s | 4.4 MB 00:00 2026-03-06T13:48:48.015 INFO:teuthology.orchestra.run.vm00.stdout:(101/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 1.4 MB/s | 25 kB 00:00 2026-03-06T13:48:48.018 INFO:teuthology.orchestra.run.vm00.stdout:(102/150): liboath-2.6.12-1.el9.x86_64.rpm 17 MB/s | 49 kB 00:00 2026-03-06T13:48:48.019 INFO:teuthology.orchestra.run.vm00.stdout:(103/150): libunwind-1.6.2-1.el9.x86_64.rpm 17 MB/s | 67 kB 00:00 2026-03-06T13:48:48.022 INFO:teuthology.orchestra.run.vm00.stdout:(104/150): luarocks-3.9.2-5.el9.noarch.rpm 41 MB/s | 151 kB 00:00 2026-03-06T13:48:48.037 INFO:teuthology.orchestra.run.vm00.stdout:(105/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 48 MB/s | 838 kB 00:00 2026-03-06T13:48:48.039 INFO:teuthology.orchestra.run.vm00.stdout:(106/150): python3-asyncssh-2.13.2-5.el9.noarch 33 MB/s | 548 kB 00:00 2026-03-06T13:48:48.040 INFO:teuthology.orchestra.run.vm00.stdout:(107/150): python3-autocommand-2.2.2-8.el9.noar 8.3 MB/s | 29 kB 00:00 2026-03-06T13:48:48.044 INFO:teuthology.orchestra.run.vm00.stdout:(108/150): python3-backports-tarfile-1.2.0-1.el 12 MB/s | 60 kB 00:00 2026-03-06T13:48:48.045 INFO:teuthology.orchestra.run.vm00.stdout:(109/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 9.0 MB/s | 43 kB 00:00 2026-03-06T13:48:48.047 INFO:teuthology.orchestra.run.vm00.stdout:(110/150): python3-cachetools-4.2.4-1.el9.noarc 10 MB/s | 32 kB 00:00 2026-03-06T13:48:48.048 INFO:teuthology.orchestra.run.vm00.stdout:(111/150): python3-certifi-2023.05.07-4.el9.noa 6.1 MB/s | 14 kB 00:00 2026-03-06T13:48:48.055 INFO:teuthology.orchestra.run.vm00.stdout:(112/150): python3-cheroot-10.0.1-4.el9.noarch. 23 MB/s | 173 kB 00:00 2026-03-06T13:48:48.056 INFO:teuthology.orchestra.run.vm00.stdout:(113/150): python3-cherrypy-18.6.1-2.el9.noarch 41 MB/s | 358 kB 00:00 2026-03-06T13:48:48.065 INFO:teuthology.orchestra.run.vm00.stdout:(114/150): python3-google-auth-2.45.0-1.el9.noa 25 MB/s | 254 kB 00:00 2026-03-06T13:48:48.103 INFO:teuthology.orchestra.run.vm00.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 3.8 MB/s | 144 kB 00:00 2026-03-06T13:48:48.116 INFO:teuthology.orchestra.run.vm00.stdout:(116/150): python3-grpcio-1.46.7-10.el9.x86_64. 34 MB/s | 2.0 MB 00:00 2026-03-06T13:48:48.117 INFO:teuthology.orchestra.run.vm00.stdout:(117/150): python3-influxdb-5.3.1-1.el9.noarch. 9.9 MB/s | 139 kB 00:00 2026-03-06T13:48:48.119 INFO:teuthology.orchestra.run.vm00.stdout:(118/150): python3-isodate-0.6.1-3.el9.noarch.r 21 MB/s | 56 kB 00:00 2026-03-06T13:48:48.119 INFO:teuthology.orchestra.run.vm00.stdout:(119/150): python3-jaraco-8.2.1-3.el9.noarch.rp 4.8 MB/s | 11 kB 00:00 2026-03-06T13:48:48.121 INFO:teuthology.orchestra.run.vm00.stdout:(120/150): python3-jaraco-classes-3.2.1-5.el9.n 7.8 MB/s | 18 kB 00:00 2026-03-06T13:48:48.122 INFO:teuthology.orchestra.run.vm00.stdout:(121/150): python3-jaraco-collections-3.0.0-8.e 9.5 MB/s | 23 kB 00:00 2026-03-06T13:48:48.124 INFO:teuthology.orchestra.run.vm00.stdout:(122/150): python3-jaraco-context-6.0.1-3.el9.n 8.6 MB/s | 20 kB 00:00 2026-03-06T13:48:48.125 INFO:teuthology.orchestra.run.vm00.stdout:(123/150): python3-jaraco-functools-3.5.0-2.el9 6.6 MB/s | 19 kB 00:00 2026-03-06T13:48:48.126 INFO:teuthology.orchestra.run.vm00.stdout:(124/150): python3-jaraco-text-4.0.0-2.el9.noar 11 MB/s | 26 kB 00:00 2026-03-06T13:48:48.130 INFO:teuthology.orchestra.run.vm00.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 13 MB/s | 46 kB 00:00 2026-03-06T13:48:48.134 INFO:teuthology.orchestra.run.vm00.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 22 MB/s | 79 kB 00:00 2026-03-06T13:48:48.138 INFO:teuthology.orchestra.run.vm00.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 21 MB/s | 86 kB 00:00 2026-03-06T13:48:48.141 INFO:teuthology.orchestra.run.vm00.stdout:(128/150): python3-kubernetes-26.1.0-3.el9.noar 66 MB/s | 1.0 MB 00:00 2026-03-06T13:48:48.141 INFO:teuthology.orchestra.run.vm00.stdout:(129/150): python3-natsort-7.1.1-5.el9.noarch.r 17 MB/s | 58 kB 00:00 2026-03-06T13:48:48.146 INFO:teuthology.orchestra.run.vm00.stdout:(130/150): python3-portend-3.1.0-2.el9.noarch.r 3.3 MB/s | 16 kB 00:00 2026-03-06T13:48:48.150 INFO:teuthology.orchestra.run.vm00.stdout:(131/150): python3-pecan-1.4.2-3.el9.noarch.rpm 30 MB/s | 272 kB 00:00 2026-03-06T13:48:48.152 INFO:teuthology.orchestra.run.vm00.stdout:(132/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 15 MB/s | 90 kB 00:00 2026-03-06T13:48:48.156 INFO:teuthology.orchestra.run.vm00.stdout:(133/150): python3-repoze-lru-0.7-16.el9.noarch 5.1 MB/s | 31 kB 00:00 2026-03-06T13:48:48.163 INFO:teuthology.orchestra.run.vm00.stdout:(134/150): python3-rsa-4.9-2.el9.noarch.rpm 7.9 MB/s | 59 kB 00:00 2026-03-06T13:48:48.166 INFO:teuthology.orchestra.run.vm00.stdout:(135/150): python3-routes-2.5.1-5.el9.noarch.rp 14 MB/s | 188 kB 00:00 2026-03-06T13:48:48.167 INFO:teuthology.orchestra.run.vm00.stdout:(136/150): python3-saml-1.16.0-1.el9.noarch.rpm 36 MB/s | 125 kB 00:00 2026-03-06T13:48:48.169 INFO:teuthology.orchestra.run.vm00.stdout:(137/150): python3-tempora-5.0.0-2.el9.noarch.r 11 MB/s | 36 kB 00:00 2026-03-06T13:48:48.170 INFO:teuthology.orchestra.run.vm00.stdout:(138/150): python3-typing-extensions-4.15.0-1.e 28 MB/s | 86 kB 00:00 2026-03-06T13:48:48.179 INFO:teuthology.orchestra.run.vm00.stdout:(139/150): python3-websocket-client-1.2.3-2.el9 9.6 MB/s | 90 kB 00:00 2026-03-06T13:48:48.180 INFO:teuthology.orchestra.run.vm01.stdout:(3/150): ceph-base-19.2.3-47.gc24117fd552.el9.c 8.8 MB/s | 5.5 MB 00:00 2026-03-06T13:48:48.182 INFO:teuthology.orchestra.run.vm00.stdout:(140/150): python3-webob-1.8.8-2.el9.noarch.rpm 18 MB/s | 230 kB 00:00 2026-03-06T13:48:48.186 INFO:teuthology.orchestra.run.vm00.stdout:(141/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 11 MB/s | 48 kB 00:00 2026-03-06T13:48:48.188 INFO:teuthology.orchestra.run.vm00.stdout:(142/150): python3-werkzeug-2.0.3-3.el9.1.noarc 49 MB/s | 427 kB 00:00 2026-03-06T13:48:48.189 INFO:teuthology.orchestra.run.vm00.stdout:(143/150): python3-xmltodict-0.12.0-15.el9.noar 8.7 MB/s | 22 kB 00:00 2026-03-06T13:48:48.191 INFO:teuthology.orchestra.run.vm00.stdout:(144/150): python3-zc-lockfile-2.0-10.el9.noarc 7.3 MB/s | 20 kB 00:00 2026-03-06T13:48:48.193 INFO:teuthology.orchestra.run.vm00.stdout:(145/150): re2-20211101-20.el9.x86_64.rpm 40 MB/s | 191 kB 00:00 2026-03-06T13:48:48.195 INFO:teuthology.orchestra.run.vm00.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 47 MB/s | 206 kB 00:00 2026-03-06T13:48:48.224 INFO:teuthology.orchestra.run.vm00.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 53 MB/s | 1.6 MB 00:00 2026-03-06T13:48:48.232 INFO:teuthology.orchestra.run.vm01.stdout:(4/150): ceph-immutable-object-cache-19.2.3-47. 449 kB/s | 145 kB 00:00 2026-03-06T13:48:48.326 INFO:teuthology.orchestra.run.vm01.stdout:(5/150): ceph-mgr-19.2.3-47.gc24117fd552.el9.cl 11 MB/s | 1.1 MB 00:00 2026-03-06T13:48:48.349 INFO:teuthology.orchestra.run.vm01.stdout:(6/150): ceph-mds-19.2.3-47.gc24117fd552.el9.cl 14 MB/s | 2.4 MB 00:00 2026-03-06T13:48:48.632 INFO:teuthology.orchestra.run.vm01.stdout:(7/150): ceph-mon-19.2.3-47.gc24117fd552.el9.cl 15 MB/s | 4.7 MB 00:00 2026-03-06T13:48:48.692 INFO:teuthology.orchestra.run.vm00.stdout:(148/150): librados2-19.2.3-47.gc24117fd552.el9 6.9 MB/s | 3.4 MB 00:00 2026-03-06T13:48:48.794 INFO:teuthology.orchestra.run.vm00.stdout:(149/150): librbd1-19.2.3-47.gc24117fd552.el9.c 5.6 MB/s | 3.2 MB 00:00 2026-03-06T13:48:48.956 INFO:teuthology.orchestra.run.vm01.stdout:(8/150): ceph-common-19.2.3-47.gc24117fd552.el9 16 MB/s | 22 MB 00:01 2026-03-06T13:48:49.003 INFO:teuthology.orchestra.run.vm01.stdout:(9/150): ceph-osd-19.2.3-47.gc24117fd552.el9.cl 26 MB/s | 17 MB 00:00 2026-03-06T13:48:49.004 INFO:teuthology.orchestra.run.vm01.stdout:(10/150): ceph-selinux-19.2.3-47.gc24117fd552.e 519 kB/s | 25 kB 00:00 2026-03-06T13:48:49.071 INFO:teuthology.orchestra.run.vm01.stdout:(11/150): libcephfs-devel-19.2.3-47.gc24117fd55 505 kB/s | 34 kB 00:00 2026-03-06T13:48:49.198 INFO:teuthology.orchestra.run.vm01.stdout:(12/150): libcephfs2-19.2.3-47.gc24117fd552.el9 7.7 MB/s | 998 kB 00:00 2026-03-06T13:48:49.240 INFO:teuthology.orchestra.run.vm01.stdout:(13/150): libcephsqlite-19.2.3-47.gc24117fd552. 3.8 MB/s | 163 kB 00:00 2026-03-06T13:48:49.278 INFO:teuthology.orchestra.run.vm01.stdout:(14/150): librados-devel-19.2.3-47.gc24117fd552 3.2 MB/s | 127 kB 00:00 2026-03-06T13:48:49.348 INFO:teuthology.orchestra.run.vm01.stdout:(15/150): ceph-radosgw-19.2.3-47.gc24117fd552.e 15 MB/s | 11 MB 00:00 2026-03-06T13:48:49.355 INFO:teuthology.orchestra.run.vm01.stdout:(16/150): libradosstriper1-19.2.3-47.gc24117fd5 6.4 MB/s | 503 kB 00:00 2026-03-06T13:48:49.401 INFO:teuthology.orchestra.run.vm01.stdout:(17/150): python3-ceph-argparse-19.2.3-47.gc241 996 kB/s | 45 kB 00:00 2026-03-06T13:48:49.493 INFO:teuthology.orchestra.run.vm01.stdout:(18/150): python3-ceph-common-19.2.3-47.gc24117 1.5 MB/s | 142 kB 00:00 2026-03-06T13:48:49.540 INFO:teuthology.orchestra.run.vm01.stdout:(19/150): python3-cephfs-19.2.3-47.gc24117fd552 3.5 MB/s | 165 kB 00:00 2026-03-06T13:48:49.626 INFO:teuthology.orchestra.run.vm01.stdout:(20/150): python3-rados-19.2.3-47.gc24117fd552. 3.8 MB/s | 322 kB 00:00 2026-03-06T13:48:49.649 INFO:teuthology.orchestra.run.vm01.stdout:(21/150): librgw2-19.2.3-47.gc24117fd552.el9.cl 18 MB/s | 5.4 MB 00:00 2026-03-06T13:48:49.662 INFO:teuthology.orchestra.run.vm01.stdout:(22/150): python3-rbd-19.2.3-47.gc24117fd552.el 7.8 MB/s | 303 kB 00:00 2026-03-06T13:48:49.682 INFO:teuthology.orchestra.run.vm01.stdout:(23/150): python3-rgw-19.2.3-47.gc24117fd552.el 3.0 MB/s | 100 kB 00:00 2026-03-06T13:48:49.696 INFO:teuthology.orchestra.run.vm01.stdout:(24/150): rbd-fuse-19.2.3-47.gc24117fd552.el9.c 2.5 MB/s | 85 kB 00:00 2026-03-06T13:48:49.742 INFO:teuthology.orchestra.run.vm01.stdout:(25/150): rbd-nbd-19.2.3-47.gc24117fd552.el9.cl 3.6 MB/s | 171 kB 00:00 2026-03-06T13:48:49.791 INFO:teuthology.orchestra.run.vm01.stdout:(26/150): ceph-grafana-dashboards-19.2.3-47.gc2 640 kB/s | 31 kB 00:00 2026-03-06T13:48:49.830 INFO:teuthology.orchestra.run.vm01.stdout:(27/150): ceph-mgr-cephadm-19.2.3-47.gc24117fd5 3.8 MB/s | 150 kB 00:00 2026-03-06T13:48:50.105 INFO:teuthology.orchestra.run.vm01.stdout:(28/150): ceph-test-19.2.3-47.gc24117fd552.el9. 45 MB/s | 50 MB 00:01 2026-03-06T13:48:50.158 INFO:teuthology.orchestra.run.vm01.stdout:(29/150): ceph-mgr-dashboard-19.2.3-47.gc24117f 12 MB/s | 3.8 MB 00:00 2026-03-06T13:48:50.167 INFO:teuthology.orchestra.run.vm01.stdout:(30/150): rbd-mirror-19.2.3-47.gc24117fd552.el9 6.4 MB/s | 3.1 MB 00:00 2026-03-06T13:48:50.189 INFO:teuthology.orchestra.run.vm01.stdout:(31/150): ceph-mgr-k8sevents-19.2.3-47.gc24117f 723 kB/s | 22 kB 00:00 2026-03-06T13:48:50.222 INFO:teuthology.orchestra.run.vm01.stdout:(32/150): ceph-mgr-modules-core-19.2.3-47.gc241 4.5 MB/s | 252 kB 00:00 2026-03-06T13:48:50.223 INFO:teuthology.orchestra.run.vm01.stdout:(33/150): ceph-mgr-rook-19.2.3-47.gc24117fd552. 1.4 MB/s | 49 kB 00:00 2026-03-06T13:48:50.257 INFO:teuthology.orchestra.run.vm01.stdout:(34/150): ceph-prometheus-alerts-19.2.3-47.gc24 475 kB/s | 16 kB 00:00 2026-03-06T13:48:50.350 INFO:teuthology.orchestra.run.vm01.stdout:(35/150): cephadm-19.2.3-47.gc24117fd552.el9.cl 8.1 MB/s | 769 kB 00:00 2026-03-06T13:48:50.353 INFO:teuthology.orchestra.run.vm01.stdout:(36/150): ceph-volume-19.2.3-47.gc24117fd552.el 2.2 MB/s | 299 kB 00:00 2026-03-06T13:48:50.469 INFO:teuthology.orchestra.run.vm01.stdout:(37/150): bzip2-1.0.8-11.el9.x86_64.rpm 461 kB/s | 55 kB 00:00 2026-03-06T13:48:50.527 INFO:teuthology.orchestra.run.vm01.stdout:(38/150): ceph-mgr-diskprediction-local-19.2.3- 18 MB/s | 7.4 MB 00:00 2026-03-06T13:48:50.529 INFO:teuthology.orchestra.run.vm01.stdout:(39/150): fuse-2.9.9-17.el9.x86_64.rpm 1.3 MB/s | 80 kB 00:00 2026-03-06T13:48:50.531 INFO:teuthology.orchestra.run.vm01.stdout:(40/150): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.9 MB/s | 351 kB 00:00 2026-03-06T13:48:50.612 INFO:teuthology.orchestra.run.vm01.stdout:(41/150): libconfig-1.7.2-9.el9.x86_64.rpm 873 kB/s | 72 kB 00:00 2026-03-06T13:48:50.616 INFO:teuthology.orchestra.run.vm01.stdout:(42/150): ledmon-libs-1.1.0-3.el9.x86_64.rpm 454 kB/s | 40 kB 00:00 2026-03-06T13:48:50.647 INFO:teuthology.orchestra.run.vm01.stdout:(43/150): mailcap-2.1.49-5.el9.noarch.rpm 1.1 MB/s | 33 kB 00:00 2026-03-06T13:48:50.784 INFO:teuthology.orchestra.run.vm01.stdout:(44/150): libgfortran-11.5.0-14.el9.x86_64.rpm 3.1 MB/s | 794 kB 00:00 2026-03-06T13:48:50.791 INFO:teuthology.orchestra.run.vm01.stdout:(45/150): pciutils-3.7.0-7.el9.x86_64.rpm 648 kB/s | 93 kB 00:00 2026-03-06T13:48:50.825 INFO:teuthology.orchestra.run.vm01.stdout:(46/150): libquadmath-11.5.0-14.el9.x86_64.rpm 869 kB/s | 184 kB 00:00 2026-03-06T13:48:50.852 INFO:teuthology.orchestra.run.vm01.stdout:(47/150): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.6 MB/s | 253 kB 00:00 2026-03-06T13:48:50.918 INFO:teuthology.orchestra.run.vm00.stdout:(150/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 5.6 MB/s | 19 MB 00:03 2026-03-06T13:48:50.919 INFO:teuthology.orchestra.run.vm01.stdout:(48/150): python3-ply-3.11-14.el9.noarch.rpm 1.1 MB/s | 106 kB 00:00 2026-03-06T13:48:50.920 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:48:50.921 INFO:teuthology.orchestra.run.vm00.stdout:Total 23 MB/s | 214 MB 00:09 2026-03-06T13:48:50.936 INFO:teuthology.orchestra.run.vm01.stdout:(49/150): python3-pycparser-2.20-6.el9.noarch.r 1.6 MB/s | 135 kB 00:00 2026-03-06T13:48:50.951 INFO:teuthology.orchestra.run.vm01.stdout:(50/150): python3-requests-2.25.1-10.el9.noarch 3.8 MB/s | 126 kB 00:00 2026-03-06T13:48:50.955 INFO:teuthology.orchestra.run.vm01.stdout:(51/150): python3-cryptography-36.0.1-5.el9.x86 7.6 MB/s | 1.2 MB 00:00 2026-03-06T13:48:50.970 INFO:teuthology.orchestra.run.vm01.stdout:(52/150): python3-urllib3-1.26.5-7.el9.noarch.r 6.4 MB/s | 218 kB 00:00 2026-03-06T13:48:50.988 INFO:teuthology.orchestra.run.vm01.stdout:(53/150): unzip-6.0-59.el9.x86_64.rpm 5.4 MB/s | 182 kB 00:00 2026-03-06T13:48:51.014 INFO:teuthology.orchestra.run.vm01.stdout:(54/150): smartmontools-7.2-10.el9.x86_64.rpm 8.6 MB/s | 556 kB 00:00 2026-03-06T13:48:51.030 INFO:teuthology.orchestra.run.vm01.stdout:(55/150): zip-3.0-35.el9.x86_64.rpm 4.3 MB/s | 266 kB 00:00 2026-03-06T13:48:51.100 INFO:teuthology.orchestra.run.vm01.stdout:(56/150): flexiblas-3.0.4-9.el9.x86_64.rpm 348 kB/s | 30 kB 00:00 2026-03-06T13:48:51.128 INFO:teuthology.orchestra.run.vm01.stdout:(57/150): boost-program-options-1.75.0-13.el9.x 745 kB/s | 104 kB 00:00 2026-03-06T13:48:51.129 INFO:teuthology.orchestra.run.vm01.stdout:(58/150): flexiblas-openblas-openmp-3.0.4-9.el9 517 kB/s | 15 kB 00:00 2026-03-06T13:48:51.186 INFO:teuthology.orchestra.run.vm01.stdout:(59/150): libnbd-1.20.3-4.el9.x86_64.rpm 2.8 MB/s | 164 kB 00:00 2026-03-06T13:48:51.214 INFO:teuthology.orchestra.run.vm01.stdout:(60/150): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.8 MB/s | 160 kB 00:00 2026-03-06T13:48:51.215 INFO:teuthology.orchestra.run.vm01.stdout:(61/150): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.5 MB/s | 45 kB 00:00 2026-03-06T13:48:51.248 INFO:teuthology.orchestra.run.vm01.stdout:(62/150): libstoragemgmt-1.10.1-1.el9.x86_64.rp 7.5 MB/s | 246 kB 00:00 2026-03-06T13:48:51.279 INFO:teuthology.orchestra.run.vm01.stdout:(63/150): libxslt-1.1.34-12.el9.x86_64.rpm 7.3 MB/s | 233 kB 00:00 2026-03-06T13:48:51.297 INFO:teuthology.orchestra.run.vm01.stdout:(64/150): flexiblas-netlib-3.0.4-9.el9.x86_64.r 11 MB/s | 3.0 MB 00:00 2026-03-06T13:48:51.300 INFO:teuthology.orchestra.run.vm01.stdout:(65/150): librdkafka-1.6.1-102.el9.x86_64.rpm 7.6 MB/s | 662 kB 00:00 2026-03-06T13:48:51.311 INFO:teuthology.orchestra.run.vm01.stdout:(66/150): lttng-ust-2.12.0-6.el9.x86_64.rpm 9.0 MB/s | 292 kB 00:00 2026-03-06T13:48:51.328 INFO:teuthology.orchestra.run.vm01.stdout:(67/150): lua-5.4.4-4.el9.x86_64.rpm 6.0 MB/s | 188 kB 00:00 2026-03-06T13:48:51.329 INFO:teuthology.orchestra.run.vm01.stdout:(68/150): openblas-0.3.29-1.el9.x86_64.rpm 1.4 MB/s | 42 kB 00:00 2026-03-06T13:48:51.357 INFO:teuthology.orchestra.run.vm01.stdout:(69/150): perl-Benchmark-1.23-483.el9.noarch.rp 924 kB/s | 26 kB 00:00 2026-03-06T13:48:51.361 INFO:teuthology.orchestra.run.vm01.stdout:(70/150): perl-Test-Harness-3.42-461.el9.noarch 9.1 MB/s | 295 kB 00:00 2026-03-06T13:48:51.547 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:48:51.601 INFO:teuthology.orchestra.run.vm01.stdout:(71/150): protobuf-3.14.0-17.el9.x86_64.rpm 4.1 MB/s | 1.0 MB 00:00 2026-03-06T13:48:51.606 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:48:51.606 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:48:51.634 INFO:teuthology.orchestra.run.vm01.stdout:(72/150): python3-devel-3.9.25-3.el9.x86_64.rpm 7.3 MB/s | 244 kB 00:00 2026-03-06T13:48:51.672 INFO:teuthology.orchestra.run.vm01.stdout:(73/150): python3-babel-2.9.1-2.el9.noarch.rpm 19 MB/s | 6.0 MB 00:00 2026-03-06T13:48:51.673 INFO:teuthology.orchestra.run.vm01.stdout:(74/150): python3-jinja2-2.11.3-8.el9.noarch.rp 6.2 MB/s | 249 kB 00:00 2026-03-06T13:48:51.702 INFO:teuthology.orchestra.run.vm01.stdout:(75/150): openblas-openmp-0.3.29-1.el9.x86_64.r 14 MB/s | 5.3 MB 00:00 2026-03-06T13:48:51.703 INFO:teuthology.orchestra.run.vm01.stdout:(76/150): python3-libstoragemgmt-1.10.1-1.el9.x 5.5 MB/s | 177 kB 00:00 2026-03-06T13:48:51.734 INFO:teuthology.orchestra.run.vm01.stdout:(77/150): python3-mako-1.1.4-6.el9.noarch.rpm 5.3 MB/s | 172 kB 00:00 2026-03-06T13:48:51.735 INFO:teuthology.orchestra.run.vm01.stdout:(78/150): python3-markupsafe-1.1.1-12.el9.x86_6 1.1 MB/s | 35 kB 00:00 2026-03-06T13:48:51.778 INFO:teuthology.orchestra.run.vm01.stdout:(79/150): python3-numpy-f2py-1.23.5-2.el9.x86_6 10 MB/s | 442 kB 00:00 2026-03-06T13:48:51.807 INFO:teuthology.orchestra.run.vm01.stdout:(80/150): python3-packaging-20.9-5.el9.noarch.r 2.6 MB/s | 77 kB 00:00 2026-03-06T13:48:51.839 INFO:teuthology.orchestra.run.vm01.stdout:(81/150): python3-protobuf-3.14.0-17.el9.noarch 8.3 MB/s | 267 kB 00:00 2026-03-06T13:48:51.870 INFO:teuthology.orchestra.run.vm01.stdout:(82/150): python3-pyasn1-0.4.8-7.el9.noarch.rpm 5.0 MB/s | 157 kB 00:00 2026-03-06T13:48:51.902 INFO:teuthology.orchestra.run.vm01.stdout:(83/150): python3-pyasn1-modules-0.4.8-7.el9.no 8.6 MB/s | 277 kB 00:00 2026-03-06T13:48:51.932 INFO:teuthology.orchestra.run.vm01.stdout:(84/150): python3-requests-oauthlib-1.3.0-12.el 1.8 MB/s | 54 kB 00:00 2026-03-06T13:48:51.982 INFO:teuthology.orchestra.run.vm01.stdout:(85/150): python3-numpy-1.23.5-2.el9.x86_64.rpm 25 MB/s | 6.1 MB 00:00 2026-03-06T13:48:52.000 INFO:teuthology.orchestra.run.vm01.stdout:(86/150): python3-lxml-4.6.5-3.el9.x86_64.rpm 3.8 MB/s | 1.2 MB 00:00 2026-03-06T13:48:52.011 INFO:teuthology.orchestra.run.vm01.stdout:(87/150): python3-toml-0.10.2-6.el9.noarch.rpm 1.4 MB/s | 42 kB 00:00 2026-03-06T13:48:52.031 INFO:teuthology.orchestra.run.vm01.stdout:(88/150): qatlib-25.08.0-2.el9.x86_64.rpm 7.5 MB/s | 240 kB 00:00 2026-03-06T13:48:52.039 INFO:teuthology.orchestra.run.vm01.stdout:(89/150): qatlib-service-25.08.0-2.el9.x86_64.r 1.3 MB/s | 37 kB 00:00 2026-03-06T13:48:52.061 INFO:teuthology.orchestra.run.vm01.stdout:(90/150): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.2 MB/s | 66 kB 00:00 2026-03-06T13:48:52.071 INFO:teuthology.orchestra.run.vm01.stdout:(91/150): socat-1.7.4.1-8.el9.x86_64.rpm 9.4 MB/s | 303 kB 00:00 2026-03-06T13:48:52.092 INFO:teuthology.orchestra.run.vm01.stdout:(92/150): xmlsec1-1.2.29-13.el9.x86_64.rpm 6.0 MB/s | 189 kB 00:00 2026-03-06T13:48:52.101 INFO:teuthology.orchestra.run.vm01.stdout:(93/150): xmlsec1-openssl-1.2.29-13.el9.x86_64. 3.0 MB/s | 90 kB 00:00 2026-03-06T13:48:52.121 INFO:teuthology.orchestra.run.vm01.stdout:(94/150): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.1 MB/s | 64 kB 00:00 2026-03-06T13:48:52.284 INFO:teuthology.orchestra.run.vm01.stdout:(95/150): lua-devel-5.4.4-4.el9.x86_64.rpm 122 kB/s | 22 kB 00:00 2026-03-06T13:48:52.303 INFO:teuthology.orchestra.run.vm01.stdout:(96/150): abseil-cpp-20211102.0-4.el9.x86_64.rp 29 MB/s | 551 kB 00:00 2026-03-06T13:48:52.309 INFO:teuthology.orchestra.run.vm01.stdout:(97/150): gperftools-libs-2.9.1-3.el9.x86_64.rp 49 MB/s | 308 kB 00:00 2026-03-06T13:48:52.311 INFO:teuthology.orchestra.run.vm01.stdout:(98/150): grpc-data-1.46.7-10.el9.noarch.rpm 9.7 MB/s | 19 kB 00:00 2026-03-06T13:48:52.378 INFO:teuthology.orchestra.run.vm01.stdout:(99/150): libarrow-9.0.0-15.el9.x86_64.rpm 66 MB/s | 4.4 MB 00:00 2026-03-06T13:48:52.381 INFO:teuthology.orchestra.run.vm01.stdout:(100/150): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-06T13:48:52.383 INFO:teuthology.orchestra.run.vm01.stdout:(101/150): liboath-2.6.12-1.el9.x86_64.rpm 19 MB/s | 49 kB 00:00 2026-03-06T13:48:52.386 INFO:teuthology.orchestra.run.vm01.stdout:(102/150): libunwind-1.6.2-1.el9.x86_64.rpm 26 MB/s | 67 kB 00:00 2026-03-06T13:48:52.390 INFO:teuthology.orchestra.run.vm01.stdout:(103/150): luarocks-3.9.2-5.el9.noarch.rpm 43 MB/s | 151 kB 00:00 2026-03-06T13:48:52.401 INFO:teuthology.orchestra.run.vm01.stdout:(104/150): protobuf-compiler-3.14.0-17.el9.x86_ 3.0 MB/s | 862 kB 00:00 2026-03-06T13:48:52.405 INFO:teuthology.orchestra.run.vm01.stdout:(105/150): parquet-libs-9.0.0-15.el9.x86_64.rpm 53 MB/s | 838 kB 00:00 2026-03-06T13:48:52.408 INFO:teuthology.orchestra.run.vm01.stdout:(106/150): python3-autocommand-2.2.2-8.el9.noar 14 MB/s | 29 kB 00:00 2026-03-06T13:48:52.410 INFO:teuthology.orchestra.run.vm01.stdout:(107/150): python3-backports-tarfile-1.2.0-1.el 24 MB/s | 60 kB 00:00 2026-03-06T13:48:52.413 INFO:teuthology.orchestra.run.vm01.stdout:(108/150): python3-bcrypt-3.2.2-1.el9.x86_64.rp 19 MB/s | 43 kB 00:00 2026-03-06T13:48:52.416 INFO:teuthology.orchestra.run.vm01.stdout:(109/150): python3-cachetools-4.2.4-1.el9.noarc 11 MB/s | 32 kB 00:00 2026-03-06T13:48:52.418 INFO:teuthology.orchestra.run.vm01.stdout:(110/150): python3-certifi-2023.05.07-4.el9.noa 6.7 MB/s | 14 kB 00:00 2026-03-06T13:48:52.422 INFO:teuthology.orchestra.run.vm01.stdout:(111/150): python3-cheroot-10.0.1-4.el9.noarch. 45 MB/s | 173 kB 00:00 2026-03-06T13:48:52.423 INFO:teuthology.orchestra.run.vm01.stdout:(112/150): python3-asyncssh-2.13.2-5.el9.noarch 24 MB/s | 548 kB 00:00 2026-03-06T13:48:52.428 INFO:teuthology.orchestra.run.vm01.stdout:(113/150): python3-cherrypy-18.6.1-2.el9.noarch 55 MB/s | 358 kB 00:00 2026-03-06T13:48:52.440 INFO:teuthology.orchestra.run.vm01.stdout:(114/150): python3-google-auth-2.45.0-1.el9.noa 15 MB/s | 254 kB 00:00 2026-03-06T13:48:52.454 INFO:teuthology.orchestra.run.vm01.stdout:(115/150): python3-grpcio-tools-1.46.7-10.el9.x 10 MB/s | 144 kB 00:00 2026-03-06T13:48:52.462 INFO:teuthology.orchestra.run.vm01.stdout:(116/150): python3-grpcio-1.46.7-10.el9.x86_64. 61 MB/s | 2.0 MB 00:00 2026-03-06T13:48:52.463 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:48:52.463 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:48:52.464 INFO:teuthology.orchestra.run.vm01.stdout:(117/150): python3-influxdb-5.3.1-1.el9.noarch. 13 MB/s | 139 kB 00:00 2026-03-06T13:48:52.465 INFO:teuthology.orchestra.run.vm01.stdout:(118/150): python3-isodate-0.6.1-3.el9.noarch.r 19 MB/s | 56 kB 00:00 2026-03-06T13:48:52.466 INFO:teuthology.orchestra.run.vm01.stdout:(119/150): python3-jaraco-8.2.1-3.el9.noarch.rp 5.6 MB/s | 11 kB 00:00 2026-03-06T13:48:52.467 INFO:teuthology.orchestra.run.vm01.stdout:(120/150): python3-jaraco-classes-3.2.1-5.el9.n 7.9 MB/s | 18 kB 00:00 2026-03-06T13:48:52.468 INFO:teuthology.orchestra.run.vm01.stdout:(121/150): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-06T13:48:52.470 INFO:teuthology.orchestra.run.vm01.stdout:(122/150): python3-jaraco-context-6.0.1-3.el9.n 7.3 MB/s | 20 kB 00:00 2026-03-06T13:48:52.471 INFO:teuthology.orchestra.run.vm01.stdout:(123/150): python3-jaraco-functools-3.5.0-2.el9 9.0 MB/s | 19 kB 00:00 2026-03-06T13:48:52.472 INFO:teuthology.orchestra.run.vm01.stdout:(124/150): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-06T13:48:52.475 INFO:teuthology.orchestra.run.vm01.stdout:(125/150): python3-logutils-0.3.5-21.el9.noarch 18 MB/s | 46 kB 00:00 2026-03-06T13:48:52.479 INFO:teuthology.orchestra.run.vm01.stdout:(126/150): python3-more-itertools-8.12.0-2.el9. 21 MB/s | 79 kB 00:00 2026-03-06T13:48:52.482 INFO:teuthology.orchestra.run.vm01.stdout:(127/150): python3-msgpack-1.0.3-2.el9.x86_64.r 30 MB/s | 86 kB 00:00 2026-03-06T13:48:52.485 INFO:teuthology.orchestra.run.vm01.stdout:(128/150): python3-natsort-7.1.1-5.el9.noarch.r 18 MB/s | 58 kB 00:00 2026-03-06T13:48:52.492 INFO:teuthology.orchestra.run.vm01.stdout:(129/150): python3-kubernetes-26.1.0-3.el9.noar 48 MB/s | 1.0 MB 00:00 2026-03-06T13:48:52.494 INFO:teuthology.orchestra.run.vm01.stdout:(130/150): python3-pecan-1.4.2-3.el9.noarch.rpm 33 MB/s | 272 kB 00:00 2026-03-06T13:48:52.494 INFO:teuthology.orchestra.run.vm01.stdout:(131/150): python3-portend-3.1.0-2.el9.noarch.r 8.5 MB/s | 16 kB 00:00 2026-03-06T13:48:52.497 INFO:teuthology.orchestra.run.vm01.stdout:(132/150): python3-pyOpenSSL-21.0.0-1.el9.noarc 31 MB/s | 90 kB 00:00 2026-03-06T13:48:52.497 INFO:teuthology.orchestra.run.vm01.stdout:(133/150): python3-repoze-lru-0.7-16.el9.noarch 11 MB/s | 31 kB 00:00 2026-03-06T13:48:52.500 INFO:teuthology.orchestra.run.vm01.stdout:(134/150): python3-routes-2.5.1-5.el9.noarch.rp 51 MB/s | 188 kB 00:00 2026-03-06T13:48:52.501 INFO:teuthology.orchestra.run.vm01.stdout:(135/150): python3-rsa-4.9-2.el9.noarch.rpm 15 MB/s | 59 kB 00:00 2026-03-06T13:48:52.504 INFO:teuthology.orchestra.run.vm01.stdout:(136/150): python3-saml-1.16.0-1.el9.noarch.rpm 32 MB/s | 125 kB 00:00 2026-03-06T13:48:52.505 INFO:teuthology.orchestra.run.vm01.stdout:(137/150): python3-tempora-5.0.0-2.el9.noarch.r 9.1 MB/s | 36 kB 00:00 2026-03-06T13:48:52.507 INFO:teuthology.orchestra.run.vm01.stdout:(138/150): python3-typing-extensions-4.15.0-1.e 30 MB/s | 86 kB 00:00 2026-03-06T13:48:52.511 INFO:teuthology.orchestra.run.vm01.stdout:(139/150): python3-websocket-client-1.2.3-2.el9 28 MB/s | 90 kB 00:00 2026-03-06T13:48:52.512 INFO:teuthology.orchestra.run.vm01.stdout:(140/150): python3-webob-1.8.8-2.el9.noarch.rpm 35 MB/s | 230 kB 00:00 2026-03-06T13:48:52.517 INFO:teuthology.orchestra.run.vm01.stdout:(141/150): python3-werkzeug-2.0.3-3.el9.1.noarc 65 MB/s | 427 kB 00:00 2026-03-06T13:48:52.518 INFO:teuthology.orchestra.run.vm01.stdout:(142/150): python3-xmlsec-1.3.13-1.el9.x86_64.r 7.8 MB/s | 48 kB 00:00 2026-03-06T13:48:52.520 INFO:teuthology.orchestra.run.vm01.stdout:(143/150): python3-xmltodict-0.12.0-15.el9.noar 7.3 MB/s | 22 kB 00:00 2026-03-06T13:48:52.521 INFO:teuthology.orchestra.run.vm01.stdout:(144/150): python3-zc-lockfile-2.0-10.el9.noarc 6.5 MB/s | 20 kB 00:00 2026-03-06T13:48:52.526 INFO:teuthology.orchestra.run.vm01.stdout:(145/150): re2-20211101-20.el9.x86_64.rpm 37 MB/s | 191 kB 00:00 2026-03-06T13:48:52.534 INFO:teuthology.orchestra.run.vm01.stdout:(146/150): s3cmd-2.4.0-1.el9.noarch.rpm 16 MB/s | 206 kB 00:00 2026-03-06T13:48:52.549 INFO:teuthology.orchestra.run.vm01.stdout:(147/150): thrift-0.15.0-4.el9.x86_64.rpm 68 MB/s | 1.6 MB 00:00 2026-03-06T13:48:52.794 INFO:teuthology.orchestra.run.vm01.stdout:(148/150): python3-scipy-1.9.3-2.el9.x86_64.rpm 22 MB/s | 19 MB 00:00 2026-03-06T13:48:52.860 INFO:teuthology.orchestra.run.vm01.stdout:(149/150): librados2-19.2.3-47.gc24117fd552.el9 11 MB/s | 3.4 MB 00:00 2026-03-06T13:48:52.925 INFO:teuthology.orchestra.run.vm01.stdout:(150/150): librbd1-19.2.3-47.gc24117fd552.el9.c 8.4 MB/s | 3.2 MB 00:00 2026-03-06T13:48:52.928 INFO:teuthology.orchestra.run.vm01.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:48:52.928 INFO:teuthology.orchestra.run.vm01.stdout:Total 32 MB/s | 214 MB 00:06 2026-03-06T13:48:53.411 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:48:53.427 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:48:53.439 INFO:teuthology.orchestra.run.vm00.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:48:53.555 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:48:53.620 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:48:53.620 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:48:53.623 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:48:53.626 INFO:teuthology.orchestra.run.vm00.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:48:53.690 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:48:53.692 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:48:53.723 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:48:53.729 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:48:53.740 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:48:53.745 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:48:53.747 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:48:53.752 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:48:53.794 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:48:53.804 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:48:53.813 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:48:53.815 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:48:53.852 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:48:53.853 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:48:53.869 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:48:53.905 INFO:teuthology.orchestra.run.vm00.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:48:53.942 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:48:53.950 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:48:53.975 INFO:teuthology.orchestra.run.vm00.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:48:53.991 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:48:53.999 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:48:54.010 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:48:54.017 INFO:teuthology.orchestra.run.vm00.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:48:54.022 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:48:54.027 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:48:54.058 INFO:teuthology.orchestra.run.vm00.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:48:54.075 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:48:54.081 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:48:54.089 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:48:54.091 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:48:54.126 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:48:54.134 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:48:54.145 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:48:54.161 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:48:54.170 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:48:54.203 INFO:teuthology.orchestra.run.vm00.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:48:54.209 INFO:teuthology.orchestra.run.vm00.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:48:54.218 INFO:teuthology.orchestra.run.vm00.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:48:54.248 INFO:teuthology.orchestra.run.vm00.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:48:54.310 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:48:54.327 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:48:54.331 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:48:54.337 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:48:54.344 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:48:54.349 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:48:54.359 INFO:teuthology.orchestra.run.vm00.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:48:54.365 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:48:54.370 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:48:54.388 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:48:54.414 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:48:54.422 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:48:54.429 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:48:54.445 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:48:54.459 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:48:54.469 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:48:54.474 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:48:54.474 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:48:54.494 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:48:54.507 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:48:54.576 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:48:54.593 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:48:54.614 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:48:54.621 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:48:54.631 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:48:54.683 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:48:55.096 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:48:55.114 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:48:55.121 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:48:55.130 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:48:55.134 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:48:55.143 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:48:55.147 INFO:teuthology.orchestra.run.vm00.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:48:55.149 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:48:55.182 INFO:teuthology.orchestra.run.vm00.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:48:55.237 INFO:teuthology.orchestra.run.vm00.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:48:55.250 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:48:55.262 INFO:teuthology.orchestra.run.vm00.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:48:55.267 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:48:55.275 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:48:55.280 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:48:55.289 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:48:55.294 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:48:55.328 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:48:55.343 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:48:55.388 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:48:55.388 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:48:55.403 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/152 2026-03-06T13:48:55.415 INFO:teuthology.orchestra.run.vm01.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/152 2026-03-06T13:48:55.578 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/152 2026-03-06T13:48:55.581 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:48:55.638 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:48:55.640 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:48:55.667 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:48:55.668 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 5/152 2026-03-06T13:48:55.675 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 6/152 2026-03-06T13:48:55.685 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 7/152 2026-03-06T13:48:55.688 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 8/152 2026-03-06T13:48:55.691 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 9/152 2026-03-06T13:48:55.695 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 10/152 2026-03-06T13:48:55.699 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:48:55.703 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:48:55.707 INFO:teuthology.orchestra.run.vm00.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:48:55.733 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 11/152 2026-03-06T13:48:55.741 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 12/152 2026-03-06T13:48:55.751 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 13/152 2026-03-06T13:48:55.752 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:48:55.770 INFO:teuthology.orchestra.run.vm00.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:48:55.773 INFO:teuthology.orchestra.run.vm00.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:48:55.788 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 14/152 2026-03-06T13:48:55.789 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:48:55.798 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:48:55.804 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 15/152 2026-03-06T13:48:55.837 INFO:teuthology.orchestra.run.vm01.stdout: Installing : re2-1:20211101-20.el9.x86_64 16/152 2026-03-06T13:48:55.873 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 17/152 2026-03-06T13:48:55.878 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 18/152 2026-03-06T13:48:55.902 INFO:teuthology.orchestra.run.vm01.stdout: Installing : liboath-2.6.12-1.el9.x86_64 19/152 2026-03-06T13:48:55.915 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 20/152 2026-03-06T13:48:55.923 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-packaging-20.9-5.el9.noarch 21/152 2026-03-06T13:48:55.932 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 22/152 2026-03-06T13:48:55.939 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 23/152 2026-03-06T13:48:55.942 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/152 2026-03-06T13:48:55.948 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 25/152 2026-03-06T13:48:55.974 INFO:teuthology.orchestra.run.vm01.stdout: Installing : unzip-6.0-59.el9.x86_64 26/152 2026-03-06T13:48:55.990 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 27/152 2026-03-06T13:48:55.994 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 28/152 2026-03-06T13:48:56.002 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 29/152 2026-03-06T13:48:56.004 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 30/152 2026-03-06T13:48:56.033 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/152 2026-03-06T13:48:56.039 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 32/152 2026-03-06T13:48:56.049 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:48:56.062 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 34/152 2026-03-06T13:48:56.070 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 35/152 2026-03-06T13:48:56.098 INFO:teuthology.orchestra.run.vm01.stdout: Installing : zip-3.0-35.el9.x86_64 36/152 2026-03-06T13:48:56.104 INFO:teuthology.orchestra.run.vm01.stdout: Installing : luarocks-3.9.2-5.el9.noarch 37/152 2026-03-06T13:48:56.111 INFO:teuthology.orchestra.run.vm01.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 38/152 2026-03-06T13:48:56.140 INFO:teuthology.orchestra.run.vm01.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 39/152 2026-03-06T13:48:56.200 INFO:teuthology.orchestra.run.vm00.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:48:56.204 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 40/152 2026-03-06T13:48:56.221 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 41/152 2026-03-06T13:48:56.225 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/152 2026-03-06T13:48:56.230 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/152 2026-03-06T13:48:56.237 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/152 2026-03-06T13:48:56.242 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/152 2026-03-06T13:48:56.251 INFO:teuthology.orchestra.run.vm01.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 46/152 2026-03-06T13:48:56.257 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 47/152 2026-03-06T13:48:56.262 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 48/152 2026-03-06T13:48:56.279 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 49/152 2026-03-06T13:48:56.299 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:48:56.304 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 50/152 2026-03-06T13:48:56.312 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 51/152 2026-03-06T13:48:56.322 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 52/152 2026-03-06T13:48:56.335 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 53/152 2026-03-06T13:48:56.347 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 54/152 2026-03-06T13:48:56.356 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 55/152 2026-03-06T13:48:56.379 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 56/152 2026-03-06T13:48:56.390 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 57/152 2026-03-06T13:48:56.450 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 58/152 2026-03-06T13:48:56.466 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 59/152 2026-03-06T13:48:56.484 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 60/152 2026-03-06T13:48:56.491 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 61/152 2026-03-06T13:48:56.500 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 62/152 2026-03-06T13:48:56.546 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 63/152 2026-03-06T13:48:56.912 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 64/152 2026-03-06T13:48:56.927 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 65/152 2026-03-06T13:48:56.933 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 66/152 2026-03-06T13:48:56.940 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 67/152 2026-03-06T13:48:56.945 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 68/152 2026-03-06T13:48:56.951 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 69/152 2026-03-06T13:48:56.955 INFO:teuthology.orchestra.run.vm01.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 70/152 2026-03-06T13:48:56.956 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 71/152 2026-03-06T13:48:56.985 INFO:teuthology.orchestra.run.vm01.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 72/152 2026-03-06T13:48:57.034 INFO:teuthology.orchestra.run.vm01.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 73/152 2026-03-06T13:48:57.048 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 74/152 2026-03-06T13:48:57.054 INFO:teuthology.orchestra.run.vm01.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 75/152 2026-03-06T13:48:57.060 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 76/152 2026-03-06T13:48:57.067 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 77/152 2026-03-06T13:48:57.071 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 78/152 2026-03-06T13:48:57.080 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:48:57.080 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 79/152 2026-03-06T13:48:57.085 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 80/152 2026-03-06T13:48:57.104 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:48:57.117 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 81/152 2026-03-06T13:48:57.129 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 82/152 2026-03-06T13:48:57.170 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 83/152 2026-03-06T13:48:57.265 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:48:57.268 INFO:teuthology.orchestra.run.vm00.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:48:57.299 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:48:57.303 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:48:57.311 INFO:teuthology.orchestra.run.vm00.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:48:57.421 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 84/152 2026-03-06T13:48:57.450 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 85/152 2026-03-06T13:48:57.453 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 86/152 2026-03-06T13:48:57.457 INFO:teuthology.orchestra.run.vm01.stdout: Installing : perl-Benchmark-1.23-483.el9.noarch 87/152 2026-03-06T13:48:57.515 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-0.3.29-1.el9.x86_64 88/152 2026-03-06T13:48:57.518 INFO:teuthology.orchestra.run.vm01.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 89/152 2026-03-06T13:48:57.540 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 90/152 2026-03-06T13:48:57.576 INFO:teuthology.orchestra.run.vm00.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:48:57.579 INFO:teuthology.orchestra.run.vm00.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:48:57.598 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:48:57.601 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:48:57.904 INFO:teuthology.orchestra.run.vm01.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 91/152 2026-03-06T13:48:57.989 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 92/152 2026-03-06T13:48:58.733 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 93/152 2026-03-06T13:48:58.738 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:48:58.743 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:48:58.757 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 94/152 2026-03-06T13:48:58.771 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:48:58.775 INFO:teuthology.orchestra.run.vm00.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:48:58.788 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:48:58.788 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:48:58.788 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:58.814 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:48:58.837 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:48:58.908 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 95/152 2026-03-06T13:48:58.910 INFO:teuthology.orchestra.run.vm01.stdout: Upgrading : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:48:58.939 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:48:58.941 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 96/152 2026-03-06T13:48:58.947 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 97/152 2026-03-06T13:48:58.953 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:48:58.954 INFO:teuthology.orchestra.run.vm01.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 98/152 2026-03-06T13:48:58.990 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:48:59.028 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:48:59.089 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:48:59.099 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:48:59.104 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:48:59.110 INFO:teuthology.orchestra.run.vm00.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:48:59.114 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:48:59.116 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:48:59.132 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:48:59.200 INFO:teuthology.orchestra.run.vm01.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 99/152 2026-03-06T13:48:59.202 INFO:teuthology.orchestra.run.vm01.stdout: Installing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:48:59.221 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 100/152 2026-03-06T13:48:59.223 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 101/152 2026-03-06T13:48:59.431 INFO:teuthology.orchestra.run.vm00.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:48:59.437 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:48:59.478 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:48:59.478 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:48:59.478 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:48:59.478 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:48:59.483 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:49:00.292 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:49:00.297 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:49:00.325 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 102/152 2026-03-06T13:49:00.328 INFO:teuthology.orchestra.run.vm01.stdout: Installing : smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:49:00.340 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/152 2026-03-06T13:49:00.340 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-03-06T13:49:00.340 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:00.365 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-ply-3.11-14.el9.noarch 104/152 2026-03-06T13:49:00.384 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 105/152 2026-03-06T13:49:00.469 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 106/152 2026-03-06T13:49:00.482 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 107/152 2026-03-06T13:49:00.510 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 108/152 2026-03-06T13:49:00.545 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 109/152 2026-03-06T13:49:00.604 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 110/152 2026-03-06T13:49:00.614 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 111/152 2026-03-06T13:49:00.619 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 112/152 2026-03-06T13:49:00.625 INFO:teuthology.orchestra.run.vm01.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 113/152 2026-03-06T13:49:00.629 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 114/152 2026-03-06T13:49:00.631 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:49:00.648 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 115/152 2026-03-06T13:49:00.943 INFO:teuthology.orchestra.run.vm01.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 116/152 2026-03-06T13:49:00.993 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:49:01.031 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 117/152 2026-03-06T13:49:01.031 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-06T13:49:01.031 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-06T13:49:01.031 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:01.036 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /sys 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /proc 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /mnt 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /var/tmp 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /home 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /root 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /tmp 2026-03-06T13:49:05.566 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:05.719 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:49:05.743 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:49:05.744 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:05.744 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:49:05.744 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:49:05.744 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:49:05.744 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:06.031 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:49:06.055 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:06.108 INFO:teuthology.orchestra.run.vm00.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:49:06.192 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:49:06.334 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:06.335 INFO:teuthology.orchestra.run.vm00.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:49:06.335 INFO:teuthology.orchestra.run.vm00.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:49:06.335 INFO:teuthology.orchestra.run.vm00.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:49:06.335 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:06.398 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:06.434 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:06.434 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:49:06.434 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:06.456 INFO:teuthology.orchestra.run.vm00.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:49:06.482 INFO:teuthology.orchestra.run.vm00.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:49:06.625 INFO:teuthology.orchestra.run.vm00.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:49:06.655 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:49:06.673 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:49:06.673 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:06.673 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:49:06.673 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 118/152 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-06T13:49:07.198 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:07.316 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 119/152 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-06T13:49:07.338 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:07.435 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:49:07.456 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:07.509 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:49:07.512 INFO:teuthology.orchestra.run.vm00.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:49:07.518 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:49:07.539 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:49:07.542 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:49:07.556 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:49:07.575 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 120/152 2026-03-06T13:49:07.575 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:07.575 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:49:07.575 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:49:07.576 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-06T13:49:07.576 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:07.583 INFO:teuthology.orchestra.run.vm01.stdout: Installing : mailcap-2.1.49-5.el9.noarch 121/152 2026-03-06T13:49:07.585 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 122/152 2026-03-06T13:49:07.600 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:07.600 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'qat' with GID 994. 2026-03-06T13:49:07.600 INFO:teuthology.orchestra.run.vm01.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-06T13:49:07.600 INFO:teuthology.orchestra.run.vm01.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-06T13:49:07.600 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:07.609 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:07.633 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 123/152 2026-03-06T13:49:07.633 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-06T13:49:07.633 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:07.652 INFO:teuthology.orchestra.run.vm01.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 124/152 2026-03-06T13:49:07.677 INFO:teuthology.orchestra.run.vm01.stdout: Installing : fuse-2.9.9-17.el9.x86_64 125/152 2026-03-06T13:49:07.747 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 126/152 2026-03-06T13:49:07.751 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:49:07.764 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 127/152 2026-03-06T13:49:07.764 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:07.764 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:49:07.764 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:08.054 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:49:08.059 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:49:08.504 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 128/152 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-06T13:49:08.526 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:08.540 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:49:08.542 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:49:08.555 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:49:08.556 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:49:08.583 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:49:08.586 INFO:teuthology.orchestra.run.vm01.stdout: Installing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 129/152 2026-03-06T13:49:08.591 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 130/152 2026-03-06T13:49:08.613 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:49:08.613 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 131/152 2026-03-06T13:49:08.616 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:49:08.666 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:49:08.668 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:49:08.688 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:08.702 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:49:08.711 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:49:08.765 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:49:09.135 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 132/152 2026-03-06T13:49:09.140 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:49:09.250 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:49:09.254 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:49:09.274 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:09.285 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:49:09.303 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:49:09.303 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:09.303 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:49:09.303 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:09.450 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:49:09.470 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:09.616 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 133/152 2026-03-06T13:49:09.618 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:49:09.630 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 134/152 2026-03-06T13:49:09.632 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:49:09.688 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 135/152 2026-03-06T13:49:09.742 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 136/152 2026-03-06T13:49:09.744 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 137/152 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-06T13:49:09.764 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:09.778 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:49:09.788 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 138/152 2026-03-06T13:49:09.842 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 139/152 2026-03-06T13:49:10.315 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 140/152 2026-03-06T13:49:10.319 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 141/152 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-06T13:49:10.338 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:10.349 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:49:10.367 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 142/152 2026-03-06T13:49:10.368 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:10.368 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:49:10.368 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:10.513 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 143/152 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-06T13:49:10.533 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:11.878 INFO:teuthology.orchestra.run.vm00.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:49:11.888 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:49:11.919 INFO:teuthology.orchestra.run.vm00.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:49:11.925 INFO:teuthology.orchestra.run.vm00.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:49:11.941 INFO:teuthology.orchestra.run.vm00.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:49:11.948 INFO:teuthology.orchestra.run.vm00.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:49:11.951 INFO:teuthology.orchestra.run.vm00.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:49:11.951 INFO:teuthology.orchestra.run.vm00.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:49:11.965 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:49:11.966 INFO:teuthology.orchestra.run.vm00.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:13.024 INFO:teuthology.orchestra.run.vm01.stdout: Installing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 144/152 2026-03-06T13:49:13.050 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 145/152 2026-03-06T13:49:13.122 INFO:teuthology.orchestra.run.vm01.stdout: Installing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 146/152 2026-03-06T13:49:13.190 INFO:teuthology.orchestra.run.vm01.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/152 2026-03-06T13:49:13.206 INFO:teuthology.orchestra.run.vm01.stdout: Installing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 148/152 2026-03-06T13:49:13.214 INFO:teuthology.orchestra.run.vm01.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 149/152 2026-03-06T13:49:13.217 INFO:teuthology.orchestra.run.vm01.stdout: Installing : bzip2-1.0.8-11.el9.x86_64 150/152 2026-03-06T13:49:13.217 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:49:13.227 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:49:13.228 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:49:13.229 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:49:13.230 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:49:13.231 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:49:13.232 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:49:13.235 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 151/152 2026-03-06T13:49:13.235 INFO:teuthology.orchestra.run.vm01.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:13.331 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout:Upgraded: 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout:Installed: 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.332 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:49:13.333 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:13.334 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:49:13.420 DEBUG:teuthology.parallel:result is None 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86 4/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 7/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 8/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 9/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso. 10/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 11/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86 12/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.cly 13/152 2026-03-06T13:49:14.547 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x8 14/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 15/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clys 16/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 17/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 18/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.e 19/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 20/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clys 21/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso 22/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x 23/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x 24/152 2026-03-06T13:49:14.548 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 25/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 26/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 27/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 28/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 29/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 30/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 31/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 32/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 33/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 34/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 35/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 36/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarc 37/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : bzip2-1.0.8-11.el9.x86_64 38/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 39/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 40/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 41/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 42/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 43/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 44/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 45/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 46/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 47/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 49/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 50/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 51/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 52/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 53/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 54/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 55/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 56/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 57/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 58/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 59/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 60/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 61/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 62/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 63/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 64/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 65/152 2026-03-06T13:49:14.549 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 66/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 67/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 68/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 69/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : perl-Benchmark-1.23-483.el9.noarch 70/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 71/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 72/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 73/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 74/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 75/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 76/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 77/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 78/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 79/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 80/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 81/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 82/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 83/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 84/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 85/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 86/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 87/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 88/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 89/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 90/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 91/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 92/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 93/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 94/152 2026-03-06T13:49:14.550 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 95/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 96/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 97/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 98/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 99/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 100/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 101/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 102/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 103/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 104/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 105/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 106/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 107/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 108/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 109/152 2026-03-06T13:49:14.551 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 110/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 111/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 112/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 113/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 114/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 115/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 116/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 117/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 118/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 119/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 120/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 121/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 122/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 123/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 124/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 125/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 126/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 127/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 128/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 129/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 130/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 131/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 132/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 133/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 134/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 135/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 136/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 137/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 138/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 139/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 140/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 141/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 142/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 143/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 144/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 145/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 146/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 147/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 148/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86 149/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 150/152 2026-03-06T13:49:14.552 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 151/152 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 152/152 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout:Upgraded: 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout:Installed: 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: bzip2-1.0.8-11.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: fuse-2.9.9-17.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.653 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: perl-Benchmark-1.23-483.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:49:14.654 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: s3cmd-2.4.0-1.el9.noarch 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:49:14.655 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:49:14.739 DEBUG:teuthology.parallel:result is None 2026-03-06T13:49:14.739 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:49:14.739 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-03-06T13:49:14.739 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-06T13:49:14.739 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:49:14.739 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:49:14.766 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:49:14.766 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-06T13:49:14.813 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-06T13:49:14.813 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:49:14.813 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:49:14.835 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:49:14.896 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:49:14.896 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/daemon-helper 2026-03-06T13:49:14.917 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-06T13:49:14.977 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-06T13:49:14.977 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:49:14.977 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:49:14.999 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:49:15.060 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:49:15.060 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-06T13:49:15.082 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-06T13:49:15.143 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-06T13:49:15.143 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:49:15.143 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:49:15.164 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:49:15.225 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:49:15.225 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/usr/bin/stdin-killer 2026-03-06T13:49:15.246 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-06T13:49:15.307 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 3}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': True}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'c24117fd5525679b799527bc1bd1f1dd0a2db5e2', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3'}} 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Cluster fsid is e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.100', 'mon.b': '192.168.123.101'} 2026-03-06T13:49:15.347 INFO:tasks.cephadm:First mon is mon.a on vm00 2026-03-06T13:49:15.347 INFO:tasks.cephadm:First mgr is a 2026-03-06T13:49:15.347 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-06T13:49:15.347 DEBUG:teuthology.orchestra.run.vm00:> sudo hostname $(hostname -s) 2026-03-06T13:49:15.372 DEBUG:teuthology.orchestra.run.vm01:> sudo hostname $(hostname -s) 2026-03-06T13:49:15.394 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-06T13:49:15.394 DEBUG:teuthology.orchestra.run.vm00:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:16.499 INFO:teuthology.orchestra.run.vm00.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:49 /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:16.499 DEBUG:teuthology.orchestra.run.vm01:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:17.582 INFO:teuthology.orchestra.run.vm01.stdout:-rw-r--r--. 1 ubuntu ubuntu 787672 Mar 6 13:49 /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:17.582 DEBUG:teuthology.orchestra.run.vm00:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:17.596 DEBUG:teuthology.orchestra.run.vm01:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-06T13:49:17.613 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 on all hosts... 2026-03-06T13:49:17.613 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:49:17.638 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 pull 2026-03-06T13:49:17.928 INFO:teuthology.orchestra.run.vm00.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:49:17.953 INFO:teuthology.orchestra.run.vm01.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout:{ 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout: "repo_digests": [ 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout: ] 2026-03-06T13:49:36.883 INFO:teuthology.orchestra.run.vm01.stdout:} 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout:{ 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout: "ceph_version": "ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable)", 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout: "image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout: "repo_digests": [ 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b" 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout: ] 2026-03-06T13:49:36.914 INFO:teuthology.orchestra.run.vm00.stdout:} 2026-03-06T13:49:36.931 DEBUG:teuthology.orchestra.run.vm00:> sudo mkdir -p /etc/ceph 2026-03-06T13:49:36.954 DEBUG:teuthology.orchestra.run.vm01:> sudo mkdir -p /etc/ceph 2026-03-06T13:49:36.978 DEBUG:teuthology.orchestra.run.vm00:> sudo chmod 777 /etc/ceph 2026-03-06T13:49:37.016 DEBUG:teuthology.orchestra.run.vm01:> sudo chmod 777 /etc/ceph 2026-03-06T13:49:37.039 INFO:tasks.cephadm:Writing seed config... 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [global] mon election default strategy = 3 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = True 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-06T13:49:37.040 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-06T13:49:37.040 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:49:37.040 DEBUG:teuthology.orchestra.run.vm00:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-06T13:49:37.070 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = e1eed43a-195a-11f1-ac1a-e73a715574a2 mon election default strategy = 3 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = True [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-06T13:49:37.070 DEBUG:teuthology.orchestra.run.vm00:mon.a> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service 2026-03-06T13:49:37.112 DEBUG:teuthology.orchestra.run.vm00:mgr.a> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a.service 2026-03-06T13:49:37.154 INFO:tasks.cephadm:Bootstrapping... 2026-03-06T13:49:37.154 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 -v bootstrap --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.100 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:49:37.426 INFO:teuthology.orchestra.run.vm00.stdout:-------------------------------------------------------------------------------- 2026-03-06T13:49:37.427 INFO:teuthology.orchestra.run.vm00.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3', '-v', 'bootstrap', '--fsid', 'e1eed43a-195a-11f1-ac1a-e73a715574a2', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.100', '--skip-admin-label'] 2026-03-06T13:49:37.427 INFO:teuthology.orchestra.run.vm00.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-06T13:49:37.427 INFO:teuthology.orchestra.run.vm00.stdout:Verifying podman|docker is present... 2026-03-06T13:49:37.444 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:49:37.444 INFO:teuthology.orchestra.run.vm00.stdout:Verifying lvm2 is present... 2026-03-06T13:49:37.444 INFO:teuthology.orchestra.run.vm00.stdout:Verifying time synchronization is in place... 2026-03-06T13:49:37.450 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:49:37.450 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:49:37.455 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:49:37.455 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-06T13:49:37.460 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-06T13:49:37.464 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-06T13:49:37.464 INFO:teuthology.orchestra.run.vm00.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:49:37.464 INFO:teuthology.orchestra.run.vm00.stdout:Repeating the final host check... 2026-03-06T13:49:37.479 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 5.8.0 2026-03-06T13:49:37.479 INFO:teuthology.orchestra.run.vm00.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-06T13:49:37.479 INFO:teuthology.orchestra.run.vm00.stdout:systemctl is present 2026-03-06T13:49:37.479 INFO:teuthology.orchestra.run.vm00.stdout:lvcreate is present 2026-03-06T13:49:37.484 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-06T13:49:37.484 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-06T13:49:37.489 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-06T13:49:37.489 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout inactive 2026-03-06T13:49:37.494 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout enabled 2026-03-06T13:49:37.499 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stdout active 2026-03-06T13:49:37.499 INFO:teuthology.orchestra.run.vm00.stdout:Unit chronyd.service is enabled and running 2026-03-06T13:49:37.499 INFO:teuthology.orchestra.run.vm00.stdout:Host looks OK 2026-03-06T13:49:37.499 INFO:teuthology.orchestra.run.vm00.stdout:Cluster fsid: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:37.500 INFO:teuthology.orchestra.run.vm00.stdout:Acquiring lock 140589070609280 on /run/cephadm/e1eed43a-195a-11f1-ac1a-e73a715574a2.lock 2026-03-06T13:49:37.500 INFO:teuthology.orchestra.run.vm00.stdout:Lock 140589070609280 acquired on /run/cephadm/e1eed43a-195a-11f1-ac1a-e73a715574a2.lock 2026-03-06T13:49:37.500 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 3300 ... 2026-03-06T13:49:37.500 INFO:teuthology.orchestra.run.vm00.stdout:Verifying IP 192.168.123.100 port 6789 ... 2026-03-06T13:49:37.500 INFO:teuthology.orchestra.run.vm00.stdout:Base mon IP(s) is [192.168.123.100:3300, 192.168.123.100:6789], mon addrv is [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:49:37.503 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.100 metric 100 2026-03-06T13:49:37.503 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.100 metric 100 2026-03-06T13:49:37.505 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-06T13:49:37.505 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:0/64 scope link noprefixroute 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:Mon IP `192.168.123.100` is in CIDR network `192.168.123.0/24` 2026-03-06T13:49:37.507 INFO:teuthology.orchestra.run.vm00.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-06T13:49:37.508 INFO:teuthology.orchestra.run.vm00.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-06T13:49:37.508 INFO:teuthology.orchestra.run.vm00.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stdout 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Trying to pull harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3... 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Getting image source signatures 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Copying blob sha256:d21d4233fd3d4dd2f376e5ef084c47891c860682c1de15a9c0357cea5defbc91 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Copying config sha256:306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 2026-03-06T13:49:38.313 INFO:teuthology.orchestra.run.vm00.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-06T13:49:38.653 INFO:teuthology.orchestra.run.vm00.stdout:ceph: stdout ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:49:38.653 INFO:teuthology.orchestra.run.vm00.stdout:Ceph version: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable) 2026-03-06T13:49:38.653 INFO:teuthology.orchestra.run.vm00.stdout:Extracting ceph user uid/gid from container image... 2026-03-06T13:49:38.874 INFO:teuthology.orchestra.run.vm00.stdout:stat: stdout 167 167 2026-03-06T13:49:38.874 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial keys... 2026-03-06T13:49:39.081 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQBizappCEFQORAAplnGS73+yIr9DXShWvYi9w== 2026-03-06T13:49:39.281 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQBjzappuzhjCRAAcymWgwVNh4nPYhVsXil/eg== 2026-03-06T13:49:39.494 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph-authtool: stdout AQBjzappQzJUFRAAEcRj5XuE1fQ4ZnEoL5G/fA== 2026-03-06T13:49:39.494 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial monmap... 2026-03-06T13:49:39.716 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:monmaptool for a [v2:192.168.123.100:3300,v1:192.168.123.100:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:setting min_mon_release = quincy 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: set fsid to e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:49:39.717 INFO:teuthology.orchestra.run.vm00.stdout:Creating mon... 2026-03-06T13:49:39.960 INFO:teuthology.orchestra.run.vm00.stdout:create mon.a on 2026-03-06T13:49:40.107 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-06T13:49:40.218 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-06T13:49:40.333 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target → /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target. 2026-03-06T13:49:40.334 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target → /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target. 2026-03-06T13:49:40.468 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a 2026-03-06T13:49:40.468 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service: Unit ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service not loaded. 2026-03-06T13:49:40.594 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target.wants/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service → /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@.service. 2026-03-06T13:49:40.737 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:49:40.737 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:49:40.737 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon to start... 2026-03-06T13:49:40.737 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mon... 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout cluster: 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout id: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout services: 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.238412s) 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout data: 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout pgs: 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:mon is available 2026-03-06T13:49:41.154 INFO:teuthology.orchestra.run.vm00.stdout:Assimilating anything we can from ceph.conf... 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:49:41.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = True 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:49:41.554 INFO:teuthology.orchestra.run.vm00.stdout:Generating new minimal ceph.conf... 2026-03-06T13:49:41.933 INFO:teuthology.orchestra.run.vm00.stdout:Restarting the monitor... 2026-03-06T13:49:42.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53031]: 2026-03-06 13:49:42.154588855 +0100 CET m=+0.168176618 container died d2c6f6f1aec7a0893a4799e4c529ab80f8c2e6aeedb14b086559daef5af2b23c (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:49:42.446 INFO:teuthology.orchestra.run.vm00.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-06T13:49:42.526 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53031]: 2026-03-06 13:49:42.273919776 +0100 CET m=+0.287507539 container remove d2c6f6f1aec7a0893a4799e4c529ab80f8c2e6aeedb14b086559daef5af2b23c (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 bash[53031]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service: Deactivated successfully. 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 systemd[1]: Stopped Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 systemd[1]: Starting Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53110]: 2026-03-06 13:49:42.406651217 +0100 CET m=+0.012442058 container create 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53110]: 2026-03-06 13:49:42.435212401 +0100 CET m=+0.041003251 container init 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53110]: 2026-03-06 13:49:42.438556079 +0100 CET m=+0.044346918 container start 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 bash[53110]: 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 podman[53110]: 2026-03-06 13:49:42.40159811 +0100 CET m=+0.007388960 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 systemd[1]: Started Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 6 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: pidfile_write: ignore empty --pid-file 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: load: jerasure load: lrc 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Git sha 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: DB SUMMARY 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: DB Session ID: VLDQJVH56STPQHE7XFT6 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 87077 ; 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.env: 0x55ca88cb0ca0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.info_log: 0x55ca8b4d6320 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.statistics: (nil) 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.use_fsync: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.db_log_dir: 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.wal_dir: 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:49:42.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.write_buffer_manager: 0x55ca8b4db900 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.unordered_write: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.row_cache: None 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.wal_filter: None 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.wal_compression: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_open_files: -1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Compression algorithms supported: 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kZSTD supported: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.merge_operator: 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_filter: None 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:49:42.528 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55ca8b4d63e0) 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: index_type: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_index_type: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: index_shortening: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: checksum: 4 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: no_block_cache: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache: 0x55ca8b4fb1f0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_options: 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: capacity : 536870912 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: num_shard_bits : 4 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: strict_capacity_limit : 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_compressed: (nil) 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: persistent_cache: (nil) 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size: 4096 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size_deviation: 10 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_restart_interval: 16 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: index_block_restart_interval: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: metadata_block_size: 4096 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: partition_filters: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: use_delta_encoding: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: filter_policy: bloomfilter 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: whole_key_filtering: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: verify_compression: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: format_version: 5 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_index_compression: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: block_align: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: max_auto_readahead_size: 262144 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: prepopulate_block_cache: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression: NoCompression 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.num_levels: 7 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:49:42.529 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.ttl: 2592000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 757bbabe-44c3-4b47-8da1-607c9041035c 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801382462795, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801382471765, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 84038, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 245, "table_properties": {"data_size": 82204, "index_size": 223, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 581, "raw_key_size": 10134, "raw_average_key_size": 47, "raw_value_size": 76399, "raw_average_value_size": 360, "num_data_blocks": 10, "num_entries": 212, "num_filter_entries": 212, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801382, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "757bbabe-44c3-4b47-8da1-607c9041035c", "db_session_id": "VLDQJVH56STPQHE7XFT6", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801382471821, "job": 1, "event": "recovery_finished"} 2026-03-06T13:49:42.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-06T13:49:42.842 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: monmap epoch 1 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: last_changed 2026-03-06T12:49:39.578527+0000 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: min_mon_release 19 (squid) 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: election_strategy: 1 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: fsmap 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: osdmap e1: 0 total, 0 up, 0 in 2026-03-06T13:49:42.843 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:42 vm00 ceph-mon[53145]: mgrmap e1: no daemons active 2026-03-06T13:49:42.853 INFO:teuthology.orchestra.run.vm00.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-06T13:49:42.855 INFO:teuthology.orchestra.run.vm00.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:49:42.855 INFO:teuthology.orchestra.run.vm00.stdout:Creating mgr... 2026-03-06T13:49:42.855 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-06T13:49:42.855 INFO:teuthology.orchestra.run.vm00.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-06T13:49:42.991 INFO:teuthology.orchestra.run.vm00.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a 2026-03-06T13:49:42.991 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Failed to reset failed state of unit ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a.service: Unit ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a.service not loaded. 2026-03-06T13:49:43.102 INFO:teuthology.orchestra.run.vm00.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2.target.wants/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a.service → /etc/systemd/system/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@.service. 2026-03-06T13:49:43.123 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 systemd[1]: Starting Ceph mgr.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to enable service . firewalld.service is not available 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr to start... 2026-03-06T13:49:43.262 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr... 2026-03-06T13:49:43.382 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 podman[53371]: 2026-03-06 13:49:43.205785321 +0100 CET m=+0.013686186 container create eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:49:43.383 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 podman[53371]: 2026-03-06 13:49:43.245715551 +0100 CET m=+0.053616416 container init eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:49:43.383 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 podman[53371]: 2026-03-06 13:49:43.249571808 +0100 CET m=+0.057472673 container start eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:49:43.383 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 bash[53371]: eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b 2026-03-06T13:49:43.383 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 podman[53371]: 2026-03-06 13:49:43.199754994 +0100 CET m=+0.007655859 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 2026-03-06T13:49:43.383 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 systemd[1]: Started Ceph mgr.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:49:43.682 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:43.488+0000 7f381d50c100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:49:43.682 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:43 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:43.619+0000 7f381d50c100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:43.698 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:43.699 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:43.700 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (1/15)... 2026-03-06T13:49:43.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:43 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2900111664' entity='client.admin' 2026-03-06T13:49:43.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:43 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3057635126' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:44.959 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:44 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:44.666+0000 7f381d50c100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:49:45.745 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:45 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:45.480+0000 7f381d50c100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:49:45.745 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:45 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:45.594+0000 7f381d50c100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:49:46.029 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:45 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:45.835+0000 7f381d50c100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.125 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:46.126 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (2/15)... 2026-03-06T13:49:46.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:46 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1637993454' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:47.860 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:47 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:47.545+0000 7f381d50c100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:49:47.861 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:47 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:47.857+0000 7f381d50c100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:49:48.155 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:47 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:47.984+0000 7f381d50c100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:49:48.156 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:48.095+0000 7f381d50c100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:49:48.411 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:48.240+0000 7f381d50c100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:49:48.411 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:48.358+0000 7f381d50c100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:48.540 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:48.541 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (3/15)... 2026-03-06T13:49:48.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2505887584' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:49.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:48.848+0000 7f381d50c100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:49:49.209 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:48 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:48.990+0000 7f381d50c100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:49:49.958 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:49 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:49.659+0000 7f381d50c100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:49:50.832 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:50 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:50.647+0000 7f381d50c100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:49:50.832 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:50 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:50.766+0000 7f381d50c100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:50.966 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 8, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:50.967 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (4/15)... 2026-03-06T13:49:51.150 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:50 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2309748064' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:51.150 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:50 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:50.900+0000 7f381d50c100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:49:51.150 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:51 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:51.147+0000 7f381d50c100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:49:51.458 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:51 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:51.261+0000 7f381d50c100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:49:51.855 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:51 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:51.540+0000 7f381d50c100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:49:51.855 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:51 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:51.852+0000 7f381d50c100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:49:52.188 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:52.185+0000 7f381d50c100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:49:52.458 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:52.298+0000 7f381d50c100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: Activating manager daemon a 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: mgrmap e2: a(active, starting, since 0.0041336s) 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: Manager daemon a is now available 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:49:52.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:52 vm00 ceph-mon[53145]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-06T13:49:53.358 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:53.359 INFO:teuthology.orchestra.run.vm00.stdout:mgr not available, waiting (5/15)... 2026-03-06T13:49:53.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:53 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/113121884' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:53.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:53 vm00 ceph-mon[53145]: mgrmap e3: a(active, since 1.00814s) 2026-03-06T13:49:55.695 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:55 vm00 ceph-mon[53145]: mgrmap e4: a(active, since 2s) 2026-03-06T13:49:55.845 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:55.845 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:55.845 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsid": "e1eed43a-195a-11f1-ac1a-e73a715574a2", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "health": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 0 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "a" 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "quorum_age": 13, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "btime": "2026-03-06T12:49:40:762524+0000", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "restful" 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ], 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "modified": "2026-03-06T12:49:40.763226+0000", 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout }, 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:55.846 INFO:teuthology.orchestra.run.vm00.stdout:mgr is available 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [global] 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout fsid = e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.100:3300,v1:192.168.123.100:6789] 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-06T13:49:56.295 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout [osd] 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-06T13:49:56.296 INFO:teuthology.orchestra.run.vm00.stdout:Enabling cephadm module... 2026-03-06T13:49:56.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:56 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3291888316' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:49:56.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:56 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3929330293' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T13:49:56.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:56 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3929330293' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T13:49:57.629 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:57 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: ignoring --setuser ceph since I am not root 2026-03-06T13:49:57.629 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:57 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: ignoring --setgroup ceph since I am not root 2026-03-06T13:49:57.629 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:57 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3612735148' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T13:49:57.889 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:57 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:57.643+0000 7f6df743d100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:49:57.889 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:57 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:57.770+0000 7f6df743d100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-06T13:49:58.017 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 5... 2026-03-06T13:49:58.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:58 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3612735148' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T13:49:58.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:58 vm00 ceph-mon[53145]: mgrmap e5: a(active, since 5s) 2026-03-06T13:49:58.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:49:58 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/454440810' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:49:59.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:58 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:58.812+0000 7f6df743d100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:49:59.959 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:59 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:59.609+0000 7f6df743d100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:49:59.959 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:59 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:59.725+0000 7f6df743d100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:50:00.458 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:49:59 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:49:59.956+0000 7f6df743d100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:50:01.945 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:01 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:01.632+0000 7f6df743d100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:50:02.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:01 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:01.942+0000 7f6df743d100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:02.209 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:02 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:02.069+0000 7f6df743d100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:50:02.209 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:02 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:02.179+0000 7f6df743d100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:02.708 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:02 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:02.307+0000 7f6df743d100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:50:02.709 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:02 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:02.419+0000 7f6df743d100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:50:03.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:02 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:02.895+0000 7f6df743d100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:50:03.209 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:03 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:03.040+0000 7f6df743d100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:04.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:03 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:03.716+0000 7f6df743d100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:50:04.958 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:04 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:04.700+0000 7f6df743d100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:50:04.959 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:04 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:04.820+0000 7f6df743d100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:50:04.959 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:04 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:04.940+0000 7f6df743d100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:50:05.626 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:05 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:05.211+0000 7f6df743d100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:50:05.626 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:05 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:05.341+0000 7f6df743d100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:50:05.939 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:05 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:05.623+0000 7f6df743d100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:50:06.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:05 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:05.937+0000 7f6df743d100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:06.708 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:06.270+0000 7f6df743d100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:50:06.709 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:06.383+0000 7f6df743d100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: Active manager daemon a restarted 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: Activating manager daemon a 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: osdmap e2: 0 total, 0 up, 0 in 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: mgrmap e6: a(active, starting, since 0.39879s) 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:50:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:06 vm00 ceph-mon[53145]: Manager daemon a is now available 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 5 is available 2026-03-06T13:50:07.964 INFO:teuthology.orchestra.run.vm00.stdout:Setting orchestrator backend to cephadm... 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: Found migration_current of "None". Setting to last migration. 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:50:08.046 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:07 vm00 ceph-mon[53145]: mgrmap e7: a(active, since 1.40224s) 2026-03-06T13:50:08.910 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-06T13:50:08.910 INFO:teuthology.orchestra.run.vm00.stdout:Generating ssh key... 2026-03-06T13:50:09.024 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:08] ENGINE Bus STARTING 2026-03-06T13:50:09.025 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:08 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: Generating public/private rsa key pair. 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: Your identification has been saved in /tmp/tmp_ol9q0rj/key 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: Your public key has been saved in /tmp/tmp_ol9q0rj/key.pub 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: The key fingerprint is: 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: SHA256:qaHyawwuLhQwnJxdirqL6OIO2BF0PEH+Fmcdm0jfUAo ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: The key's randomart image is: 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: +---[RSA 3072]----+ 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |o =o=o E o.. | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |o*.++ . = B | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |.o..... + * . | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |.. . . + . | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |. o + S | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |.+.. o o | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |=.oo. . | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |B..oo | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: |@* .o. | 2026-03-06T13:50:09.282 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:09 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: +----[SHA256]-----+ 2026-03-06T13:50:09.873 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkC8+nEi4g4Q01eMjrCalRBWopKHlMqiXmNwY3bczVbX9DDeQNxDkS0AjI6/dz7ElU4UcDst7nGmL76lvP2DmwSfGcsH+Q0ZwV82sXdIM5ENmVB+Vp6F4hIjHiBiRIbGmS1BltreFkd4W3ILHCr19BJNj9coAB/9Y0yvHv5Otr0VX/IM6S4X9nUX00LgRBWdwDp0JRZMwheiXa6KmoE3TP9JiXnoyZXI5OpWN/GTFeu+JnJOslkmVlLL++uYNfyOqgDMaiGLDX+M9SplECrTbaKttGy9lF2YuXCrNu4rb2qybkRAErEWxWErkxckz6Uvo2C9sOqNoR4CMRlDd6+NzLVBegdDOKGpIniuPXHC9tXqsWSBs3h1LkcTHjf8toj4Ngh6FeeyJ+/CPyxJf9AQ2Z4fjzahNGvu6IvwRl6zSOGNbC93kzNhyFhOMUrb9A3IHvA6Id+A6705vk9nOimqhlfFdKDJ4Ln+thA1ao76wrTrdMH1TZos45IEpdQ2f/QPE= ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:09.874 INFO:teuthology.orchestra.run.vm00.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:50:09.874 INFO:teuthology.orchestra.run.vm00.stdout:Adding key to root@localhost authorized_keys... 2026-03-06T13:50:09.874 INFO:teuthology.orchestra.run.vm00.stdout:Adding host vm00... 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:08] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:08] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:08] ENGINE Bus STARTED 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:08] ENGINE Client ('192.168.123.100', 48430) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:10.557 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: Generating ssh key... 2026-03-06T13:50:10.558 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:10.558 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:10.558 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:10 vm00 ceph-mon[53145]: mgrmap e8: a(active, since 2s) 2026-03-06T13:50:11.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:11 vm00 ceph-mon[53145]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:11.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:11 vm00 ceph-mon[53145]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:12.582 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:12 vm00 ceph-mon[53145]: Deploying cephadm binary to vm00 2026-03-06T13:50:12.595 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Added host 'vm00' with addr '192.168.123.100' 2026-03-06T13:50:12.595 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mon service... 2026-03-06T13:50:13.095 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-06T13:50:13.095 INFO:teuthology.orchestra.run.vm00.stdout:Deploying unmanaged mgr service... 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: Added host vm00 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: Saving service mon spec with placement count:5 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: Saving service mgr spec with placement count:2 2026-03-06T13:50:13.474 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:13 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:13.553 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-06T13:50:14.573 INFO:teuthology.orchestra.run.vm00.stdout:Enabling the dashboard module... 2026-03-06T13:50:14.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:14 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:14.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:14 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/760960951' entity='client.admin' 2026-03-06T13:50:14.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:14 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:14.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:14 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3447169154' entity='client.admin' 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3020737279' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-mon[53145]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: ignoring --setuser ceph since I am not root 2026-03-06T13:50:16.090 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:15 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: ignoring --setgroup ceph since I am not root 2026-03-06T13:50:16.388 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:16 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:16.188+0000 7f2528894100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:50:16.388 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:16 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:16.320+0000 7f2528894100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for the mgr to restart... 2026-03-06T13:50:16.535 INFO:teuthology.orchestra.run.vm00.stdout:Waiting for mgr epoch 9... 2026-03-06T13:50:16.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:16 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3020737279' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T13:50:16.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:16 vm00 ceph-mon[53145]: mgrmap e9: a(active, since 9s) 2026-03-06T13:50:16.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:16 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1033590735' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:50:17.708 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:17 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:17.426+0000 7f2528894100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:50:18.573 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:18 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:18.232+0000 7f2528894100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:50:18.573 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:18 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:18.345+0000 7f2528894100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:50:18.958 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:18 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:18.570+0000 7f2528894100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:50:20.566 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:20 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:20.252+0000 7f2528894100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:50:20.931 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:20 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:20.563+0000 7f2528894100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:20.931 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:20 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:20.690+0000 7f2528894100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:50:20.931 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:20 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:20.801+0000 7f2528894100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:21.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:20 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:20.928+0000 7f2528894100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:50:21.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:21 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:21.042+0000 7f2528894100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:50:21.958 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:21 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:21.512+0000 7f2528894100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:50:21.958 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:21 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:21.652+0000 7f2528894100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:22.708 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:22 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:22.316+0000 7f2528894100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:50:23.708 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:23 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:23.314+0000 7f2528894100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:50:23.709 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:23 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:23.434+0000 7f2528894100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:50:23.709 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:23 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:23.558+0000 7f2528894100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:50:24.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:23 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:23.808+0000 7f2528894100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:50:24.208 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:23 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:23.925+0000 7f2528894100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:50:24.533 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:24 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:24.212+0000 7f2528894100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:50:24.875 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:24 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:24.530+0000 7f2528894100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:25.147 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:24 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:24.872+0000 7f2528894100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:24 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:24.986+0000 7f2528894100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: Active manager daemon a restarted 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: Activating manager daemon a 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: osdmap e3: 0 total, 0 up, 0 in 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: mgrmap e10: a(active, starting, since 0.0076792s) 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:50:25.148 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: Manager daemon a is now available 2026-03-06T13:50:25.149 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:25.149 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:25.149 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:25 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout { 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout } 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:mgr epoch 9 is available 2026-03-06T13:50:26.141 INFO:teuthology.orchestra.run.vm00.stdout:Generating a dashboard self-signed certificate... 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-06T13:50:26.257 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:26 vm00 ceph-mon[53145]: mgrmap e11: a(active, since 1.01007s) 2026-03-06T13:50:26.659 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-06T13:50:26.659 INFO:teuthology.orchestra.run.vm00.stdout:Creating initial admin user... 2026-03-06T13:50:27.293 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$o9FvnT0bhc62XMsK6XcCEOkjaVQX8dhypvEwzW7YG//90lt8PvbDC", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772801427, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-06T13:50:27.294 INFO:teuthology.orchestra.run.vm00.stdout:Fetching dashboard port number... 2026-03-06T13:50:27.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:25] ENGINE Bus STARTING 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:26] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:26] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:26] ENGINE Bus STARTED 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: [06/Mar/2026:12:50:26] ENGINE Client ('192.168.123.100', 54044) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: Deploying daemon agent.vm00 on vm00 2026-03-06T13:50:27.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:27 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:27.727 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stdout 8443 2026-03-06T13:50:27.727 INFO:teuthology.orchestra.run.vm00.stdout:firewalld does not appear to be present 2026-03-06T13:50:27.727 INFO:teuthology.orchestra.run.vm00.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout:Ceph Dashboard is now available at: 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout: URL: https://vm00.local:8443/ 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout: User: admin 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout: Password: bldmf80p09 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:27.730 INFO:teuthology.orchestra.run.vm00.stdout:Saving cluster configuration to /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config directory 2026-03-06T13:50:28.223 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout:Or, if you are only running a single cluster on this host: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: ceph telemetry on 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout:For more information see: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:28.224 INFO:teuthology.orchestra.run.vm00.stdout:Bootstrap complete. 2026-03-06T13:50:28.255 INFO:tasks.cephadm:Fetching config... 2026-03-06T13:50:28.256 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:28.256 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-06T13:50:28.275 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-06T13:50:28.276 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:28.276 DEBUG:teuthology.orchestra.run.vm00:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-06T13:50:28.336 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-06T13:50:28.336 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:28.336 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/keyring of=/dev/stdout 2026-03-06T13:50:28.410 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-06T13:50:28.410 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:28.410 DEBUG:teuthology.orchestra.run.vm00:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-06T13:50:28.468 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-06T13:50:28.469 DEBUG:teuthology.orchestra.run.vm00:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkC8+nEi4g4Q01eMjrCalRBWopKHlMqiXmNwY3bczVbX9DDeQNxDkS0AjI6/dz7ElU4UcDst7nGmL76lvP2DmwSfGcsH+Q0ZwV82sXdIM5ENmVB+Vp6F4hIjHiBiRIbGmS1BltreFkd4W3ILHCr19BJNj9coAB/9Y0yvHv5Otr0VX/IM6S4X9nUX00LgRBWdwDp0JRZMwheiXa6KmoE3TP9JiXnoyZXI5OpWN/GTFeu+JnJOslkmVlLL++uYNfyOqgDMaiGLDX+M9SplECrTbaKttGy9lF2YuXCrNu4rb2qybkRAErEWxWErkxckz6Uvo2C9sOqNoR4CMRlDd6+NzLVBegdDOKGpIniuPXHC9tXqsWSBs3h1LkcTHjf8toj4Ngh6FeeyJ+/CPyxJf9AQ2Z4fjzahNGvu6IvwRl6zSOGNbC93kzNhyFhOMUrb9A3IHvA6Id+A6705vk9nOimqhlfFdKDJ4Ln+thA1ao76wrTrdMH1TZos45IEpdQ2f/QPE= ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:50:28.578 INFO:teuthology.orchestra.run.vm00.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkC8+nEi4g4Q01eMjrCalRBWopKHlMqiXmNwY3bczVbX9DDeQNxDkS0AjI6/dz7ElU4UcDst7nGmL76lvP2DmwSfGcsH+Q0ZwV82sXdIM5ENmVB+Vp6F4hIjHiBiRIbGmS1BltreFkd4W3ILHCr19BJNj9coAB/9Y0yvHv5Otr0VX/IM6S4X9nUX00LgRBWdwDp0JRZMwheiXa6KmoE3TP9JiXnoyZXI5OpWN/GTFeu+JnJOslkmVlLL++uYNfyOqgDMaiGLDX+M9SplECrTbaKttGy9lF2YuXCrNu4rb2qybkRAErEWxWErkxckz6Uvo2C9sOqNoR4CMRlDd6+NzLVBegdDOKGpIniuPXHC9tXqsWSBs3h1LkcTHjf8toj4Ngh6FeeyJ+/CPyxJf9AQ2Z4fjzahNGvu6IvwRl6zSOGNbC93kzNhyFhOMUrb9A3IHvA6Id+A6705vk9nOimqhlfFdKDJ4Ln+thA1ao76wrTrdMH1TZos45IEpdQ2f/QPE= ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:28.596 DEBUG:teuthology.orchestra.run.vm01:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkC8+nEi4g4Q01eMjrCalRBWopKHlMqiXmNwY3bczVbX9DDeQNxDkS0AjI6/dz7ElU4UcDst7nGmL76lvP2DmwSfGcsH+Q0ZwV82sXdIM5ENmVB+Vp6F4hIjHiBiRIbGmS1BltreFkd4W3ILHCr19BJNj9coAB/9Y0yvHv5Otr0VX/IM6S4X9nUX00LgRBWdwDp0JRZMwheiXa6KmoE3TP9JiXnoyZXI5OpWN/GTFeu+JnJOslkmVlLL++uYNfyOqgDMaiGLDX+M9SplECrTbaKttGy9lF2YuXCrNu4rb2qybkRAErEWxWErkxckz6Uvo2C9sOqNoR4CMRlDd6+NzLVBegdDOKGpIniuPXHC9tXqsWSBs3h1LkcTHjf8toj4Ngh6FeeyJ+/CPyxJf9AQ2Z4fjzahNGvu6IvwRl6zSOGNbC93kzNhyFhOMUrb9A3IHvA6Id+A6705vk9nOimqhlfFdKDJ4Ln+thA1ao76wrTrdMH1TZos45IEpdQ2f/QPE= ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-06T13:50:28.627 INFO:teuthology.orchestra.run.vm01.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCkC8+nEi4g4Q01eMjrCalRBWopKHlMqiXmNwY3bczVbX9DDeQNxDkS0AjI6/dz7ElU4UcDst7nGmL76lvP2DmwSfGcsH+Q0ZwV82sXdIM5ENmVB+Vp6F4hIjHiBiRIbGmS1BltreFkd4W3ILHCr19BJNj9coAB/9Y0yvHv5Otr0VX/IM6S4X9nUX00LgRBWdwDp0JRZMwheiXa6KmoE3TP9JiXnoyZXI5OpWN/GTFeu+JnJOslkmVlLL++uYNfyOqgDMaiGLDX+M9SplECrTbaKttGy9lF2YuXCrNu4rb2qybkRAErEWxWErkxckz6Uvo2C9sOqNoR4CMRlDd6+NzLVBegdDOKGpIniuPXHC9tXqsWSBs3h1LkcTHjf8toj4Ngh6FeeyJ+/CPyxJf9AQ2Z4fjzahNGvu6IvwRl6zSOGNbC93kzNhyFhOMUrb9A3IHvA6Id+A6705vk9nOimqhlfFdKDJ4Ln+thA1ao76wrTrdMH1TZos45IEpdQ2f/QPE= ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:28.637 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-06T13:50:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:28 vm00 ceph-mon[53145]: from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:28 vm00 ceph-mon[53145]: mgrmap e12: a(active, since 2s) 2026-03-06T13:50:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:28 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1056240438' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T13:50:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:28 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2831159342' entity='client.admin' 2026-03-06T13:50:29.152 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:29.797 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-06T13:50:29.797 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-06T13:50:30.029 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1282724305' entity='client.admin' 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:30.030 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:30.196 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:30.816 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm01 2026-03-06T13:50:30.817 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:50:30.817 DEBUG:teuthology.orchestra.run.vm01:> dd of=/etc/ceph/ceph.conf 2026-03-06T13:50:30.832 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:50:30.832 DEBUG:teuthology.orchestra.run.vm01:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:30.887 INFO:tasks.cephadm:Adding host vm01 to orchestrator... 2026-03-06T13:50:30.887 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch host add vm01 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:50:31.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:31 vm00 ceph-mon[53145]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:31.378 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:32.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.client.admin.keyring 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: mgrmap e13: a(active, since 6s) 2026-03-06T13:50:32.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:32 vm00 ceph-mon[53145]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:33.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:33 vm00 ceph-mon[53145]: Deploying cephadm binary to vm01 2026-03-06T13:50:33.922 INFO:teuthology.orchestra.run.vm00.stdout:Added host 'vm01' with addr '192.168.123.101' 2026-03-06T13:50:34.073 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch host ls --format=json 2026-03-06T13:50:34.383 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:34.718 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:34.718 INFO:teuthology.orchestra.run.vm00.stdout:[{"addr": "192.168.123.100", "hostname": "vm00", "labels": [], "status": ""}, {"addr": "192.168.123.101", "hostname": "vm01", "labels": [], "status": ""}] 2026-03-06T13:50:34.897 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-06T13:50:34.898 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd crush tunables default 2026-03-06T13:50:35.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: Added host vm01 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:35.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:34 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:35.217 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:36.004 INFO:teuthology.orchestra.run.vm00.stderr:adjusted tunables profile to default 2026-03-06T13:50:36.158 INFO:tasks.cephadm:Adding mon.a on vm00 2026-03-06T13:50:36.158 INFO:tasks.cephadm:Adding mon.b on vm01 2026-03-06T13:50:36.159 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch apply mon '2;vm00:192.168.123.100=a;vm01:192.168.123.101=b' 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.client.admin.keyring 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1361094006' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:36.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:35 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:36.550 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:36.905 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mon update... 2026-03-06T13:50:37.067 DEBUG:teuthology.orchestra.run.vm01:mon.b> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b.service 2026-03-06T13:50:37.068 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T13:50:37.068 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph mon dump -f json 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: Deploying daemon agent.vm01 on vm01 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1361094006' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:37.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:36 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:37.454 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:37.873 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:50:37.873 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":1,"fsid":"e1eed43a-195a-11f1-ac1a-e73a715574a2","modified":"2026-03-06T12:49:39.578527Z","created":"2026-03-06T12:49:39.578527Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-06T13:50:37.873 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 1 2026-03-06T13:50:38.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:37 vm00 ceph-mon[53145]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm00:192.168.123.100=a;vm01:192.168.123.101=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:38.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:37 vm00 ceph-mon[53145]: Saving service mon spec with placement vm00:192.168.123.100=a;vm01:192.168.123.101=b;count:2 2026-03-06T13:50:38.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:37 vm00 ceph-mon[53145]: Deploying daemon mon.b on vm01 2026-03-06T13:50:38.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:37 vm00 ceph-mon[53145]: from='client.? 192.168.123.101:0/1357298022' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:39.054 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-06T13:50:39.054 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph mon dump -f json 2026-03-06T13:50:39.485 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.b/config 2026-03-06T13:50:39.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:39 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:39.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:39 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:39.974 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:50:39.974 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:50:39.974 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:50:39.975 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.ttl: 2592000 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:50:40.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7a8e4116-c05f-4095-bcc3-657bdde48fef 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801439973883, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801439978077, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801439, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7a8e4116-c05f-4095-bcc3-657bdde48fef", "db_session_id": "0Y26ETS7XSE7WDOW3G2I", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801439978133, "job": 1, "event": "recovery_finished"} 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x555c3a1d2e00 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: DB pointer 0x555c3a2ea000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: mon.b does not exist in monmap, will attempt to join an existing cluster 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: using public_addr v2:192.168.123.101:0/0 -> [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: starting mon.b rank -1 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: mon.b@-1(???) e0 preinit fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:39 vm01 ceph-mon[53512]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: ** DB Stats ** 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.4 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.4 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.4 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.4 0.00 0.00 1 0.004 0 0 0.0 0.0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache BinnedLRUCache@0x555c3a1d11f0#7 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 4e-06 secs_since: 0 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).mds e1 new map 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).mds e1 print_map 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: e1 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: btime 2026-03-06T12:49:40:762524+0000 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:50:40.293 INFO:journalctl@ceph.mon.b.vm01.stdout: legacy client fscid: -1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout: No filesystems configured 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mkfs e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: monmap epoch 1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: last_changed 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: min_mon_release 19 (squid) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: election_strategy: 1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: fsmap 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: osdmap e1: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e1: no daemons active 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/154133962' entity='client.admin' cmd=[{"prefix": "status"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1010217902' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1010217902' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: monmap epoch 1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: last_changed 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: min_mon_release 19 (squid) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: election_strategy: 1 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: fsmap 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: osdmap e1: 0 total, 0 up, 0 in 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e1: no daemons active 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2900111664' entity='client.admin' 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3057635126' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1637993454' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2505887584' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2309748064' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Activating manager daemon a 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e2: a(active, starting, since 0.0041336s) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Manager daemon a is now available 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14100 192.168.123.100:0/3604282492' entity='mgr.a' 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/113121884' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e3: a(active, since 1.00814s) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e4: a(active, since 2s) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3291888316' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3929330293' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3929330293' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3612735148' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3612735148' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e5: a(active, since 5s) 2026-03-06T13:50:40.294 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/454440810' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Active manager daemon a restarted 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Activating manager daemon a 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: osdmap e2: 0 total, 0 up, 0 in 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e6: a(active, starting, since 0.39879s) 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Manager daemon a is now available 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Found migration_current of "None". Setting to last migration. 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e7: a(active, since 1.40224s) 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:08] ENGINE Bus STARTING 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:08] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:08] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:08] ENGINE Bus STARTED 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:08] ENGINE Client ('192.168.123.100', 48430) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Generating ssh key... 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e8: a(active, since 2s) 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm00", "addr": "192.168.123.100", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Deploying cephadm binary to vm00 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Added host vm00 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Saving service mon spec with placement count:5 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Saving service mgr spec with placement count:2 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/760960951' entity='client.admin' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3447169154' entity='client.admin' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3020737279' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14124 192.168.123.100:0/2504422203' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]': finished 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3020737279' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e9: a(active, since 9s) 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1033590735' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Active manager daemon a restarted 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Activating manager daemon a 2026-03-06T13:50:40.295 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: osdmap e3: 0 total, 0 up, 0 in 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e10: a(active, starting, since 0.0076792s) 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Manager daemon a is now available 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm00", "caps": []}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e11: a(active, since 1.01007s) 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:25] ENGINE Bus STARTING 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14160 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:26] ENGINE Serving on http://192.168.123.100:8765 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:26] ENGINE Serving on https://192.168.123.100:7150 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:26] ENGINE Bus STARTED 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: [06/Mar/2026:12:50:26] ENGINE Client ('192.168.123.100', 54044) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Deploying daemon agent.vm00 on vm00 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e12: a(active, since 2s) 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1056240438' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2831159342' entity='client.admin' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1282724305' entity='client.admin' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm00:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.client.admin.keyring 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.296 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mgrmap e13: a(active, since 6s) 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm01", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Deploying cephadm binary to vm01 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Added host vm01 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm01:/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.client.admin.keyring 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "client.agent.vm01", "caps": []}]': finished 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1361094006' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Deploying daemon agent.vm01 on vm01 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1361094006' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm00:192.168.123.100=a;vm01:192.168.123.101=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Saving service mon spec with placement vm00:192.168.123.100=a;vm01:192.168.123.101=b;count:2 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: Deploying daemon mon.b on vm01 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='client.? 192.168.123.101:0/1357298022' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:40.297 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:40 vm01 ceph-mon[53512]: mon.b@-1(synchronizing).paxosservice(auth 1..5) refresh upgraded, format 0 -> 3 2026-03-06T13:50:45.029 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:50:45.029 INFO:teuthology.orchestra.run.vm01.stdout:{"epoch":2,"fsid":"e1eed43a-195a-11f1-ac1a-e73a715574a2","modified":"2026-03-06T12:50:40.008473Z","created":"2026-03-06T12:49:39.578527Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:3300","nonce":0},{"type":"v1","addr":"192.168.123.100:6789","nonce":0}]},"addr":"192.168.123.100:6789/0","public_addr":"192.168.123.100:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:3300","nonce":0},{"type":"v1","addr":"192.168.123.101:6789","nonce":0}]},"addr":"192.168.123.101:6789/0","public_addr":"192.168.123.101:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-06T13:50:45.029 INFO:teuthology.orchestra.run.vm01.stderr:dumped monmap epoch 2 2026-03-06T13:50:45.204 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-06T13:50:45.204 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph config generate-minimal-conf 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: mon.a calling monitor election 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='client.? 192.168.123.101:0/2641132456' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: mon.b calling monitor election 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: monmap epoch 2 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: min_mon_release 19 (squid) 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: election_strategy: 1 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: fsmap 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: mgrmap e13: a(active, since 20s) 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: overall HEALTH_OK 2026-03-06T13:50:45.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:45.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:45.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:45 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: mon.a calling monitor election 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='client.? 192.168.123.101:0/2641132456' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: mon.b calling monitor election 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: monmap epoch 2 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: min_mon_release 19 (squid) 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: election_strategy: 1 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: fsmap 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: osdmap e4: 0 total, 0 up, 0 in 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: mgrmap e13: a(active, since 20s) 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: overall HEALTH_OK 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:45 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:45.542 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:45.926 INFO:teuthology.orchestra.run.vm00.stdout:# minimal ceph.conf for e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:45.926 INFO:teuthology.orchestra.run.vm00.stdout:[global] 2026-03-06T13:50:45.926 INFO:teuthology.orchestra.run.vm00.stdout: fsid = e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:50:45.926 INFO:teuthology.orchestra.run.vm00.stdout: mon_host = [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] 2026-03-06T13:50:46.079 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-06T13:50:46.079 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:46.079 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:50:46.103 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:46.103 DEBUG:teuthology.orchestra.run.vm00:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.162 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Reconfiguring mon.a (unknown last config time)... 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Reconfiguring daemon mon.a on vm00 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: Reconfiguring daemon mon.b on vm01 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2232051494' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.163 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:46 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:46.171 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:50:46.171 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.conf 2026-03-06T13:50:46.205 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:50:46.205 DEBUG:teuthology.orchestra.run.vm01:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-06T13:50:46.278 INFO:tasks.cephadm:Adding mgr.a on vm00 2026-03-06T13:50:46.278 INFO:tasks.cephadm:Adding mgr.b on vm01 2026-03-06T13:50:46.278 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch apply mgr '2;vm00=a;vm01=b' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Updating vm00:/etc/ceph/ceph.conf 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Updating vm01:/etc/ceph/ceph.conf 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Updating vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Updating vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/config/ceph.conf 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Reconfiguring mon.a (unknown last config time)... 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:50:46.412 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Reconfiguring daemon mon.a on vm00 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Reconfiguring mon.b (monmap changed)... 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: Reconfiguring daemon mon.b on vm01 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2232051494' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:46.413 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:46 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:50:46.648 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.b/config 2026-03-06T13:50:46.988 INFO:teuthology.orchestra.run.vm01.stdout:Scheduled mgr update... 2026-03-06T13:50:47.155 DEBUG:teuthology.orchestra.run.vm01:mgr.b> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.b.service 2026-03-06T13:50:47.156 INFO:tasks.cephadm:Deploying OSDs... 2026-03-06T13:50:47.156 DEBUG:teuthology.orchestra.run.vm00:> set -ex 2026-03-06T13:50:47.156 DEBUG:teuthology.orchestra.run.vm00:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:50:47.171 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:47.172 DEBUG:teuthology.orchestra.run.vm00:> ls /dev/[sv]d? 2026-03-06T13:50:47.227 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vda 2026-03-06T13:50:47.227 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdb 2026-03-06T13:50:47.227 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdc 2026-03-06T13:50:47.227 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vdd 2026-03-06T13:50:47.227 INFO:teuthology.orchestra.run.vm00.stdout:/dev/vde 2026-03-06T13:50:47.227 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:50:47.227 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:50:47.227 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdb 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdb 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:50:31.073114374 +0100 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:48:22.920330589 +0100 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:48:22.920330589 +0100 2026-03-06T13:50:47.284 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:45:02.236000000 +0100 2026-03-06T13:50:47.284 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:50:47.346 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a[53382]: 2026-03-06T12:50:47.006+0000 7f25098a7640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-06T13:50:47.348 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:50:47.348 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:50:47.348 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.00015435 s, 3.3 MB/s 2026-03-06T13:50:47.349 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:50:47.406 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdc 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdc 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:50:31.077114375 +0100 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:48:22.937330590 +0100 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:48:22.937330590 +0100 2026-03-06T13:50:47.464 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:45:02.247000000 +0100 2026-03-06T13:50:47.465 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:50:47.528 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:50:47.528 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:50:47.528 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000174217 s, 2.9 MB/s 2026-03-06T13:50:47.529 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:50:47.585 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vdd 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vdd 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:50:31.080114375 +0100 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:48:22.918330588 +0100 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:48:22.918330588 +0100 2026-03-06T13:50:47.642 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:45:02.264000000 +0100 2026-03-06T13:50:47.642 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:50:47.657 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:47 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:47.702 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:50:47.703 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:47 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:47.704 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:50:47.704 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:50:47.704 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000116088 s, 4.4 MB/s 2026-03-06T13:50:47.705 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:50:47.761 DEBUG:teuthology.orchestra.run.vm00:> stat /dev/vde 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout: File: /dev/vde 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Access: 2026-03-06 13:50:31.083114375 +0100 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Modify: 2026-03-06 13:48:22.922330589 +0100 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout:Change: 2026-03-06 13:48:22.922330589 +0100 2026-03-06T13:50:47.818 INFO:teuthology.orchestra.run.vm00.stdout: Birth: 2026-03-06 13:45:02.327000000 +0100 2026-03-06T13:50:47.818 DEBUG:teuthology.orchestra.run.vm00:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:50:47.880 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records in 2026-03-06T13:50:47.880 INFO:teuthology.orchestra.run.vm00.stderr:1+0 records out 2026-03-06T13:50:47.880 INFO:teuthology.orchestra.run.vm00.stderr:512 bytes copied, 0.000168296 s, 3.0 MB/s 2026-03-06T13:50:47.881 DEBUG:teuthology.orchestra.run.vm00:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:50:47.938 DEBUG:teuthology.orchestra.run.vm01:> set -ex 2026-03-06T13:50:47.938 DEBUG:teuthology.orchestra.run.vm01:> dd if=/scratch_devs of=/dev/stdout 2026-03-06T13:50:47.958 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:50:47.959 DEBUG:teuthology.orchestra.run.vm01:> ls /dev/[sv]d? 2026-03-06T13:50:47.976 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:47 vm01 systemd[1]: Starting Ceph mgr.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:50:47.998 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vda 2026-03-06T13:50:47.998 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdb 2026-03-06T13:50:47.998 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdc 2026-03-06T13:50:47.998 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vdd 2026-03-06T13:50:47.998 INFO:teuthology.orchestra.run.vm01.stdout:/dev/vde 2026-03-06T13:50:47.998 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-06T13:50:47.998 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-06T13:50:47.998 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdb 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdb 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:50:41.671051286 +0100 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:48:22.102994296 +0100 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:48:22.102994296 +0100 2026-03-06T13:50:48.086 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:45:33.234000000 +0100 2026-03-06T13:50:48.086 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-06T13:50:48.122 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:50:48.122 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:50:48.122 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000178735 s, 2.9 MB/s 2026-03-06T13:50:48.123 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-06T13:50:48.197 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdc 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdc 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:50:41.674051289 +0100 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:48:22.134994305 +0100 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:48:22.134994305 +0100 2026-03-06T13:50:48.260 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:45:33.237000000 +0100 2026-03-06T13:50:48.260 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:47 vm01 podman[54889]: 2026-03-06 13:50:47.974655123 +0100 CET m=+0.015292273 container create f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 podman[54889]: 2026-03-06 13:50:48.021887419 +0100 CET m=+0.062524569 container init f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 podman[54889]: 2026-03-06 13:50:48.027606678 +0100 CET m=+0.068243828 container start f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 bash[54889]: f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 podman[54889]: 2026-03-06 13:50:47.968455025 +0100 CET m=+0.009092184 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:50:48.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 systemd[1]: Started Ceph mgr.b for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:50:48.320 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:50:48.320 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:50:48.320 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000190097 s, 2.7 MB/s 2026-03-06T13:50:48.321 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-06T13:50:48.381 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vdd 2026-03-06T13:50:48.439 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vdd 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:50:41.678051293 +0100 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:48:22.107994298 +0100 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:48:22.107994298 +0100 2026-03-06T13:50:48.440 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:45:33.253000000 +0100 2026-03-06T13:50:48.440 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-06T13:50:48.510 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:50:48.510 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:50:48.510 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000180528 s, 2.8 MB/s 2026-03-06T13:50:48.511 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-06T13:50:48.576 DEBUG:teuthology.orchestra.run.vm01:> stat /dev/vde 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout: File: /dev/vde 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Access: 2026-03-06 13:50:41.681051296 +0100 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Modify: 2026-03-06 13:48:22.113994300 +0100 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout:Change: 2026-03-06 13:48:22.113994300 +0100 2026-03-06T13:50:48.645 INFO:teuthology.orchestra.run.vm01.stdout: Birth: 2026-03-06 13:45:33.259000000 +0100 2026-03-06T13:50:48.645 DEBUG:teuthology.orchestra.run.vm01:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:48.296+0000 7f3625d00100 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:48.442+0000 7f3625d00100 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='client.14198 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm01=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: Saving service mgr spec with placement vm00=a;vm01=b;count:2 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: Deploying daemon mgr.b on vm01 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:48 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='client.14198 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm00=a;vm01=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:48.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: Saving service mgr spec with placement vm00=a;vm01=b;count:2 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: Deploying daemon mgr.b on vm01 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-06T13:50:48.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:48 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:48.711 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records in 2026-03-06T13:50:48.711 INFO:teuthology.orchestra.run.vm01.stderr:1+0 records out 2026-03-06T13:50:48.711 INFO:teuthology.orchestra.run.vm01.stderr:512 bytes copied, 0.000200155 s, 2.6 MB/s 2026-03-06T13:50:48.712 DEBUG:teuthology.orchestra.run.vm01:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-06T13:50:48.773 INFO:tasks.cephadm:Deploying osd.0 on vm00 with /dev/vde... 2026-03-06T13:50:48.773 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- lvm zap /dev/vde 2026-03-06T13:50:49.096 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: Reconfiguring daemon mgr.a on vm00 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:49.957 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-mon[53512]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: Reconfiguring mgr.a (unknown last config time)... 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: Reconfiguring daemon mgr.a on vm00 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.054 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:49 vm00 ceph-mon[53145]: Metadata not up to date on all hosts. Skipping non agent specs 2026-03-06T13:50:50.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:49 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:49.951+0000 7f3625d00100 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-06T13:50:50.560 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:50:50.590 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch daemon add osd vm00:/dev/vde 2026-03-06T13:50:50.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:50 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:50.971 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.010 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:50 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:50:51.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:51.006+0000 7f3625d00100 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-06T13:50:51.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:51.143+0000 7f3625d00100 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-06T13:50:51.791 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:51.413+0000 7f3625d00100 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-06T13:50:52.044 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:51 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:50:52.044 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:51 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:50:52.044 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:51 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:52.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:50:52.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:50:52.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:51 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:52.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:52 vm00 ceph-mon[53145]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:52.917 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:52 vm00 ceph-mon[53145]: from='client.14206 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:53.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:52 vm01 ceph-mon[53512]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:53.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:52 vm01 ceph-mon[53512]: from='client.14206 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm00:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:50:53.743 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:53.369+0000 7f3625d00100 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1885728607' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c74ee687-3963-477a-a1a7-adcd5b543db0"}]: dispatch 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1885728607' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c74ee687-3963-477a-a1a7-adcd5b543db0"}]': finished 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-mon[53512]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1935005919' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:53.740+0000 7f3625d00100 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:54.022 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:53 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:53.891+0000 7f3625d00100 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-06T13:50:54.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:53 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1885728607' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c74ee687-3963-477a-a1a7-adcd5b543db0"}]: dispatch 2026-03-06T13:50:54.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:53 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1885728607' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c74ee687-3963-477a-a1a7-adcd5b543db0"}]': finished 2026-03-06T13:50:54.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:53 vm00 ceph-mon[53145]: osdmap e5: 1 total, 0 up, 1 in 2026-03-06T13:50:54.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:53 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:50:54.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:53 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1935005919' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:50:54.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:54.019+0000 7f3625d00100 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:54.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:54.164+0000 7f3625d00100 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-06T13:50:54.791 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:54.290+0000 7f3625d00100 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-06T13:50:55.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:54 vm00 ceph-mon[53145]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:55.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-mon[53512]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:55.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:54.830+0000 7f3625d00100 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-06T13:50:55.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:54 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:54.988+0000 7f3625d00100 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-06T13:50:56.040 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:55 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:55.730+0000 7f3625d00100 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-06T13:50:57.041 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:56 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:56.727+0000 7f3625d00100 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-06T13:50:57.041 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:56 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:56.851+0000 7f3625d00100 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-06T13:50:57.041 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:56 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:56.975+0000 7f3625d00100 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-06T13:50:57.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:56 vm01 ceph-mon[53512]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:57.085 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:56 vm00 ceph-mon[53145]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:57.541 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:57.235+0000 7f3625d00100 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-06T13:50:57.541 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:57.354+0000 7f3625d00100 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-06T13:50:57.876 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:57 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:50:57.876 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:57 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:57.971 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:57.644+0000 7f3625d00100 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-06T13:50:57.971 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-06T13:50:57.971 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:50:58.291 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:57 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:57.968+0000 7f3625d00100 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-06T13:50:58.790 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:58.313+0000 7f3625d00100 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-06T13:50:58.791 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b[54913]: 2026-03-06T12:50:58.428+0000 7f3625d00100 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: Deploying daemon osd.0 on vm00 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: Standby manager daemon b started 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:50:59.128 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:58 vm00 ceph-mon[53145]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: Deploying daemon osd.0 on vm00 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: Standby manager daemon b started 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-06T13:50:59.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:58 vm01 ceph-mon[53512]: from='mgr.? 192.168.123.101:0/60597555' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-06T13:51:00.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:59 vm00 ceph-mon[53145]: mgrmap e14: a(active, since 33s), standbys: b 2026-03-06T13:51:00.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:50:59 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:51:00.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:59 vm01 ceph-mon[53512]: mgrmap e14: a(active, since 33s), standbys: b 2026-03-06T13:51:00.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:50:59 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:01.179 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:00 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:01.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:00 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:02.183 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:01 vm00 ceph-mon[53145]: from='osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:51:02.183 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:01 vm00 ceph-mon[53145]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:51:02.183 INFO:teuthology.orchestra.run.vm00.stdout:Created osd(s) 0 on host 'vm00' 2026-03-06T13:51:02.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:01 vm01 ceph-mon[53512]: from='osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:51:02.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:01 vm01 ceph-mon[53512]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-06T13:51:02.335 DEBUG:teuthology.orchestra.run.vm00:osd.0> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.0.service 2026-03-06T13:51:02.336 INFO:tasks.cephadm:Deploying osd.1 on vm01 with /dev/vde... 2026-03-06T13:51:02.336 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- lvm zap /dev/vde 2026-03-06T13:51:02.621 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.b/config 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:02 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: osdmap e6: 1 total, 0 up, 1 in 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm00", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:03.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:03.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.292 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:02 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:03.843 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:51:03.858 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch daemon add osd vm01:/dev/vde 2026-03-06T13:51:03.920 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:03 vm01 ceph-mon[53512]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-06T13:51:03.920 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:03 vm01 ceph-mon[53512]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:51:03.920 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:03 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:03.920 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:03 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:03.920 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:03 vm01 ceph-mon[53512]: from='osd.0 ' entity='osd.0' 2026-03-06T13:51:04.153 INFO:teuthology.orchestra.run.vm01.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.b/config 2026-03-06T13:51:04.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:03 vm00 ceph-mon[53145]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm00", "root=default"]}]': finished 2026-03-06T13:51:04.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:03 vm00 ceph-mon[53145]: osdmap e7: 1 total, 0 up, 1 in 2026-03-06T13:51:04.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:03 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:04.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:03 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:04.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:03 vm00 ceph-mon[53145]: from='osd.0 ' entity='osd.0' 2026-03-06T13:51:04.209 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:51:03 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0[62596]: 2026-03-06T12:51:03.762+0000 7fd542d02640 -1 osd.0 0 waiting for initial osdmap 2026-03-06T13:51:04.209 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:51:03 vm00 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0[62596]: 2026-03-06T12:51:03.767+0000 7fd53e32b640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: purged_snaps scrub starts 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: purged_snaps scrub ok 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:05.017 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963] boot 2026-03-06T13:51:05.018 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: osdmap e8: 1 total, 1 up, 1 in 2026-03-06T13:51:05.018 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:04 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:05.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: purged_snaps scrub starts 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: purged_snaps scrub ok 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: osd.0 [v2:192.168.123.100:6802/1892809963,v1:192.168.123.100:6803/1892809963] boot 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: osdmap e8: 1 total, 1 up, 1 in 2026-03-06T13:51:05.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:04 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-06T13:51:06.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: from='client.24109 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:51:06.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: from='client.? 192.168.123.101:0/3039273215' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]: dispatch 2026-03-06T13:51:06.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]: dispatch 2026-03-06T13:51:06.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]': finished 2026-03-06T13:51:06.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: osdmap e9: 2 total, 1 up, 2 in 2026-03-06T13:51:06.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:05 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:06.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: from='client.24109 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm01:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:51:06.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: from='client.? 192.168.123.101:0/3039273215' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]: dispatch 2026-03-06T13:51:06.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]: dispatch 2026-03-06T13:51:06.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "261a000e-d3a1-4785-8522-05f878b645dd"}]': finished 2026-03-06T13:51:06.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: osdmap e9: 2 total, 1 up, 2 in 2026-03-06T13:51:06.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:05 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:06 vm00 ceph-mon[53145]: pgmap v18: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:07.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:06 vm00 ceph-mon[53145]: from='client.? 192.168.123.101:0/2630050913' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:51:07.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:06 vm01 ceph-mon[53512]: pgmap v18: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:07.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:06 vm01 ceph-mon[53512]: from='client.? 192.168.123.101:0/2630050913' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-06T13:51:09.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:08 vm01 ceph-mon[53512]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:09.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:08 vm00 ceph-mon[53145]: pgmap v20: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:11.070 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.071 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:10 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:11.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:10 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:12.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:11 vm00 ceph-mon[53145]: Deploying daemon osd.1 on vm01 2026-03-06T13:51:12.222 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:11 vm01 ceph-mon[53512]: Deploying daemon osd.1 on vm01 2026-03-06T13:51:13.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:12 vm01 ceph-mon[53512]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:13.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:12 vm00 ceph-mon[53145]: pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:13.965 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:13 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:14.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:14.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:13 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:15.269 INFO:teuthology.orchestra.run.vm01.stdout:Created osd(s) 1 on host 'vm01' 2026-03-06T13:51:15.424 DEBUG:teuthology.orchestra.run.vm01:osd.1> sudo journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.1.service 2026-03-06T13:51:15.425 INFO:tasks.cephadm:Waiting for 2 OSDs to come up... 2026-03-06T13:51:15.425 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd stat -f json 2026-03-06T13:51:15.733 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:15.791 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:51:15 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:51:15.541+0000 7fe6eb8f0740 -1 osd.1 0 log_to_monitors true 2026-03-06T13:51:16.045 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:16.192 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":9,"num_osds":2,"num_up_osds":1,"osd_up_since":1772801464,"num_in_osds":2,"osd_in_since":1772801465,"num_remapped_pgs":0} 2026-03-06T13:51:16.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:16.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:51:16.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:16 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3759182910' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm01", "name": "osd_memory_target"}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-06T13:51:16.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:16 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3759182910' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:17.193 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd stat -f json 2026-03-06T13:51:17.504 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:17.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:17 vm00 ceph-mon[53145]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:51:17.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:17 vm00 ceph-mon[53145]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:51:17.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:17 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:17.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:17 vm00 ceph-mon[53145]: from='osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:51:17.527 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:17 vm00 ceph-mon[53145]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:51:17.540 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:17 vm01 ceph-mon[53512]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-06T13:51:17.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:17 vm01 ceph-mon[53512]: osdmap e10: 2 total, 1 up, 2 in 2026-03-06T13:51:17.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:17 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:17.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:17 vm01 ceph-mon[53512]: from='osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:51:17.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:17 vm01 ceph-mon[53512]: from='osd.1 ' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]: dispatch 2026-03-06T13:51:17.817 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:17.968 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":11,"num_osds":2,"num_up_osds":1,"osd_up_since":1772801464,"num_in_osds":2,"osd_in_since":1772801465,"num_remapped_pgs":0} 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1958414440' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:18.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:18 vm01 ceph-mon[53512]: from='osd.1 ' entity='osd.1' 2026-03-06T13:51:18.541 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:51:18 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:51:18.091+0000 7fe6e7871640 -1 osd.1 0 waiting for initial osdmap 2026-03-06T13:51:18.541 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:51:18 vm01 ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:51:18.096+0000 7fe6e2e9a640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm01", "root=default"]}]': finished 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: osdmap e11: 2 total, 1 up, 2 in 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1958414440' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:18.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:18 vm00 ceph-mon[53145]: from='osd.1 ' entity='osd.1' 2026-03-06T13:51:18.968 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd stat -f json 2026-03-06T13:51:19.258 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:19.540 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: purged_snaps scrub starts 2026-03-06T13:51:19.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: purged_snaps scrub ok 2026-03-06T13:51:19.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:19.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729] boot 2026-03-06T13:51:19.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: osdmap e12: 2 total, 2 up, 2 in 2026-03-06T13:51:19.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:19 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:19.563 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: purged_snaps scrub starts 2026-03-06T13:51:19.564 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: purged_snaps scrub ok 2026-03-06T13:51:19.564 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:19.564 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: osd.1 [v2:192.168.123.101:6800/3414878729,v1:192.168.123.101:6801/3414878729] boot 2026-03-06T13:51:19.564 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: osdmap e12: 2 total, 2 up, 2 in 2026-03-06T13:51:19.564 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:19 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-06T13:51:19.564 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:19.730 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":12,"num_osds":2,"num_up_osds":2,"osd_up_since":1772801479,"num_in_osds":2,"osd_in_since":1772801465,"num_remapped_pgs":0} 2026-03-06T13:51:19.731 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd dump --format=json 2026-03-06T13:51:20.021 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:20.326 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:20.326 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":13,"fsid":"e1eed43a-195a-11f1-ac1a-e73a715574a2","created":"2026-03-06T12:49:40.762969+0000","modified":"2026-03-06T12:51:20.287796+0000","last_up_change":"2026-03-06T12:51:19.094224+0000","last_in_change":"2026-03-06T12:51:05.640389+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"c74ee687-3963-477a-a1a7-adcd5b543db0","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6803","nonce":1892809963}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6805","nonce":1892809963}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6809","nonce":1892809963}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6807","nonce":1892809963}]},"public_addr":"192.168.123.100:6803/1892809963","cluster_addr":"192.168.123.100:6805/1892809963","heartbeat_back_addr":"192.168.123.100:6809/1892809963","heartbeat_front_addr":"192.168.123.100:6807/1892809963","state":["exists","up"]},{"osd":1,"uuid":"261a000e-d3a1-4785-8522-05f878b645dd","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6801","nonce":3414878729}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6803","nonce":3414878729}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6807","nonce":3414878729}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6805","nonce":3414878729}]},"public_addr":"192.168.123.101:6801/3414878729","cluster_addr":"192.168.123.101:6803/3414878729","heartbeat_back_addr":"192.168.123.101:6807/3414878729","heartbeat_front_addr":"192.168.123.101:6805/3414878729","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:02.491909+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:16.560300+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:0/2688165074":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6801/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/3991691335":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6800/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/940332754":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6801/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6800/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/2157954058":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/1082118053":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/913077137":"2026-03-07T12:50:06.385500+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:51:20.326 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:20 vm00 ceph-mon[53145]: pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:20.326 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:20 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1542623177' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:20.472 INFO:tasks.cephadm.ceph_manager.ceph:[] 2026-03-06T13:51:20.472 INFO:tasks.cephadm:Setting up client nodes... 2026-03-06T13:51:20.472 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-06T13:51:20.472 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-06T13:51:20.472 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph mgr dump --format=json 2026-03-06T13:51:20.759 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:20.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:20 vm01 ceph-mon[53512]: pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-06T13:51:20.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:20 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1542623177' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-06T13:51:21.080 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:21.244 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":14,"flags":0,"active_gid":14156,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6800","nonce":1074681661},{"type":"v1","addr":"192.168.123.100:6801","nonce":1074681661}]},"active_addr":"192.168.123.100:6801/1074681661","active_change":"2026-03-06T12:50:24.988451+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14202,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.100:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":1386693720}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":3803200679}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.100:0","nonce":3134044447}]}]} 2026-03-06T13:51:21.246 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-06T13:51:21.246 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-06T13:51:21.246 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd dump --format=json 2026-03-06T13:51:21.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:21 vm00 ceph-mon[53145]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:51:21.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:21 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2308159027' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:21.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:21 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2742363910' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:51:21.538 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:21.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:21 vm01 ceph-mon[53512]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:51:21.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:21 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2308159027' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:21.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:21 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2742363910' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-06T13:51:21.840 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:21.840 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":13,"fsid":"e1eed43a-195a-11f1-ac1a-e73a715574a2","created":"2026-03-06T12:49:40.762969+0000","modified":"2026-03-06T12:51:20.287796+0000","last_up_change":"2026-03-06T12:51:19.094224+0000","last_in_change":"2026-03-06T12:51:05.640389+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"c74ee687-3963-477a-a1a7-adcd5b543db0","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6803","nonce":1892809963}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6805","nonce":1892809963}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6809","nonce":1892809963}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6807","nonce":1892809963}]},"public_addr":"192.168.123.100:6803/1892809963","cluster_addr":"192.168.123.100:6805/1892809963","heartbeat_back_addr":"192.168.123.100:6809/1892809963","heartbeat_front_addr":"192.168.123.100:6807/1892809963","state":["exists","up"]},{"osd":1,"uuid":"261a000e-d3a1-4785-8522-05f878b645dd","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6801","nonce":3414878729}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6803","nonce":3414878729}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6807","nonce":3414878729}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6805","nonce":3414878729}]},"public_addr":"192.168.123.101:6801/3414878729","cluster_addr":"192.168.123.101:6803/3414878729","heartbeat_back_addr":"192.168.123.101:6807/3414878729","heartbeat_front_addr":"192.168.123.101:6805/3414878729","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:02.491909+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:16.560300+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:0/2688165074":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6801/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/3991691335":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6800/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/940332754":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6801/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6800/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/2157954058":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/1082118053":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/913077137":"2026-03-07T12:50:06.385500+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:51:22.009 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-06T13:51:22.009 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd dump --format=json 2026-03-06T13:51:22.298 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:22.598 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:22 vm00 ceph-mon[53145]: pgmap v31: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:22.598 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:22 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1194344877' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:22.599 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:22.599 INFO:teuthology.orchestra.run.vm00.stdout:{"epoch":13,"fsid":"e1eed43a-195a-11f1-ac1a-e73a715574a2","created":"2026-03-06T12:49:40.762969+0000","modified":"2026-03-06T12:51:20.287796+0000","last_up_change":"2026-03-06T12:51:19.094224+0000","last_in_change":"2026-03-06T12:51:05.640389+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":6,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":2,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"c74ee687-3963-477a-a1a7-adcd5b543db0","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6802","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6803","nonce":1892809963}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6804","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6805","nonce":1892809963}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6808","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6809","nonce":1892809963}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.100:6806","nonce":1892809963},{"type":"v1","addr":"192.168.123.100:6807","nonce":1892809963}]},"public_addr":"192.168.123.100:6803/1892809963","cluster_addr":"192.168.123.100:6805/1892809963","heartbeat_back_addr":"192.168.123.100:6809/1892809963","heartbeat_front_addr":"192.168.123.100:6807/1892809963","state":["exists","up"]},{"osd":1,"uuid":"261a000e-d3a1-4785-8522-05f878b645dd","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6800","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6801","nonce":3414878729}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6802","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6803","nonce":3414878729}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6806","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6807","nonce":3414878729}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.101:6804","nonce":3414878729},{"type":"v1","addr":"192.168.123.101:6805","nonce":3414878729}]},"public_addr":"192.168.123.101:6801/3414878729","cluster_addr":"192.168.123.101:6803/3414878729","heartbeat_back_addr":"192.168.123.101:6807/3414878729","heartbeat_front_addr":"192.168.123.101:6805/3414878729","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:02.491909+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-06T12:51:16.560300+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.100:0/2688165074":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6801/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/3991691335":"2026-03-07T12:50:24.988208+0000","192.168.123.100:6800/572548250":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/940332754":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6801/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:6800/161270728":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/2157954058":"2026-03-07T12:50:24.988208+0000","192.168.123.100:0/1082118053":"2026-03-07T12:50:06.385500+0000","192.168.123.100:0/913077137":"2026-03-07T12:50:06.385500+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-06T13:51:22.762 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph tell osd.0 flush_pg_stats 2026-03-06T13:51:22.762 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph tell osd.1 flush_pg_stats 2026-03-06T13:51:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:22 vm01 ceph-mon[53512]: pgmap v31: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:22.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:22 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1194344877' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:23.068 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:23.068 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:23.343 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:23 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3448698297' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:23.517 INFO:teuthology.orchestra.run.vm00.stdout:34359738373 2026-03-06T13:51:23.517 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd last-stat-seq osd.0 2026-03-06T13:51:23.519 INFO:teuthology.orchestra.run.vm00.stdout:51539607554 2026-03-06T13:51:23.519 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd last-stat-seq osd.1 2026-03-06T13:51:23.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:23 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3448698297' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-06T13:51:23.823 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:23.903 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:24.149 INFO:teuthology.orchestra.run.vm00.stdout:34359738372 2026-03-06T13:51:24.217 INFO:teuthology.orchestra.run.vm00.stdout:51539607553 2026-03-06T13:51:24.296 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738373 got 34359738372 for osd.0 2026-03-06T13:51:24.383 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607554 got 51539607553 for osd.1 2026-03-06T13:51:24.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:24 vm00 ceph-mon[53145]: pgmap v32: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:24.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:24 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/2821030955' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:51:24.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:24 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/93988189' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:51:24.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:24 vm01 ceph-mon[53512]: pgmap v32: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:24.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:24 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/2821030955' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:51:24.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:24 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/93988189' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:51:25.296 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd last-stat-seq osd.0 2026-03-06T13:51:25.383 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph osd last-stat-seq osd.1 2026-03-06T13:51:25.600 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:25.705 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:25.928 INFO:teuthology.orchestra.run.vm00.stdout:34359738374 2026-03-06T13:51:26.029 INFO:teuthology.orchestra.run.vm00.stdout:51539607555 2026-03-06T13:51:26.082 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738373 got 34359738374 for osd.0 2026-03-06T13:51:26.083 DEBUG:teuthology.parallel:result is None 2026-03-06T13:51:26.178 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607554 got 51539607555 for osd.1 2026-03-06T13:51:26.178 DEBUG:teuthology.parallel:result is None 2026-03-06T13:51:26.178 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-06T13:51:26.178 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph pg dump --format=json 2026-03-06T13:51:26.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:26 vm00 ceph-mon[53145]: pgmap v33: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:26.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:26 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/29007555' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:51:26.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:26 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/1158759053' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:51:26.464 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:26.763 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:26.763 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-06T13:51:26.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:26 vm01 ceph-mon[53512]: pgmap v33: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:26.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:26 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/29007555' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-06T13:51:26.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:26 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/1158759053' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-06T13:51:26.929 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":33,"stamp":"2026-03-06T12:51:25.004732+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":463504,"kb_used_data":224,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41471344,"statfs":{"total":42941284352,"available":42466656256,"internally_reserved":0,"allocated":229376,"data_stored":56234,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":12,"seq":51539607555,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":436552,"kb_used_data":112,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20530872,"statfs":{"total":21470642176,"available":21023612928,"internally_reserved":0,"allocated":114688,"data_stored":28117,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738374,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26952,"kb_used_data":112,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940472,"statfs":{"total":21470642176,"available":21443043328,"internally_reserved":0,"allocated":114688,"data_stored":28117,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-06T13:51:26.929 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph pg dump --format=json 2026-03-06T13:51:27.222 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:27.520 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:27.520 INFO:teuthology.orchestra.run.vm00.stderr:dumped all 2026-03-06T13:51:27.520 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:27 vm00 ceph-mon[53145]: from='client.24157 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:27.666 INFO:teuthology.orchestra.run.vm00.stdout:{"pg_ready":true,"pg_map":{"version":34,"stamp":"2026-03-06T12:51:27.004936+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":0,"num_osds":2,"num_per_pool_osds":2,"num_per_pool_omap_osds":0,"kb":41934848,"kb_used":463504,"kb_used_data":224,"kb_used_omap":3,"kb_used_meta":53628,"kb_avail":41471344,"statfs":{"total":42941284352,"available":42466656256,"internally_reserved":0,"allocated":229376,"data_stored":56234,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":3180,"internal_metadata":54915988},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"0.000000"},"pg_stats":[],"pool_stats":[],"osd_stats":[{"osd":1,"up_from":12,"seq":51539607555,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":436552,"kb_used_data":112,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20530872,"statfs":{"total":21470642176,"available":21023612928,"internally_reserved":0,"allocated":114688,"data_stored":28117,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738374,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":26952,"kb_used_data":112,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940472,"statfs":{"total":21470642176,"available":21443043328,"internally_reserved":0,"allocated":114688,"data_stored":28117,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[]}} 2026-03-06T13:51:27.666 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-06T13:51:27.666 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-06T13:51:27.666 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-06T13:51:27.666 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph health --format=json 2026-03-06T13:51:27.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:27 vm01 ceph-mon[53512]: from='client.24157 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:27.952 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:28.273 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:28.273 INFO:teuthology.orchestra.run.vm00.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-06T13:51:28.436 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-06T13:51:28.436 INFO:tasks.cephadm:Setup complete, yielding 2026-03-06T13:51:28.436 INFO:teuthology.run_tasks:Running task exec... 2026-03-06T13:51:28.438 INFO:teuthology.task.exec:Executing custom commands... 2026-03-06T13:51:28.438 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm00.local 2026-03-06T13:51:28.438 DEBUG:teuthology.orchestra.run.vm00:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-06T13:51:28.461 INFO:teuthology.task.exec:Running commands on role host.b host ubuntu@vm01.local 2026-03-06T13:51:28.461 DEBUG:teuthology.orchestra.run.vm01:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'mkdir /etc/cephadm_testing' 2026-03-06T13:51:28.484 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-06T13:51:28.488 INFO:tasks.cephadm:Applying spec(s): extra_container_args: - --cpus=2 extra_entrypoint_args: - --debug_ms 10 placement: host_pattern: '*' service_type: mon --- custom_configs: - content: "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n\ \ ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\n\ sleep infinity\n" mount_path: /root/write_thing_to_file.sh extra_container_args: - -v - /etc/cephadm_testing:/root/cephadm_testing extra_entrypoint_args: - /root/write_thing_to_file.sh - -c - testing_custom_containers - -o - /root/cephadm_testing/testing.txt placement: host_pattern: '*' service_id: foo service_type: container spec: entrypoint: bash image: quay.io/fedora/fedora:latest --- custom_configs: - content: 'set -e test -f /var/cache/bar/from.txt test -f /var/cache/bar/presized.dat echo ok > /var/cache/bar/primary.txt sleep infinity ' mount_path: /root/init_check.sh extra_entrypoint_args: - /root/init_check.sh placement: host_pattern: '*' service_id: bar service_type: container spec: dirs: - data entrypoint: bash image: quay.io/fedora/fedora:latest init_containers: - entrypoint: bash entrypoint_args: - argument: -c - argument: . /etc/os-release && echo from=$ID > /var/cache/bar/from.txt image: quay.io/centos/centos:latest volume_mounts: data: /var/cache/bar:z - entrypoint: bash entrypoint_args: - argument: -c - argument: test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat volume_mounts: data: /var/cache/bar:z volume_mounts: data: /var/cache/bar:z 2026-03-06T13:51:28.488 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch apply -i - 2026-03-06T13:51:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:28 vm00 ceph-mon[53145]: pgmap v34: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:28 vm00 ceph-mon[53145]: from='client.24161 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:28.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:28 vm00 ceph-mon[53145]: from='client.? 192.168.123.100:0/3692668731' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:51:28.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:28 vm01 ceph-mon[53512]: pgmap v34: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:28.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:28 vm01 ceph-mon[53512]: from='client.24161 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:28.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:28 vm01 ceph-mon[53512]: from='client.? 192.168.123.100:0/3692668731' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-06T13:51:28.791 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:29.110 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled mon update... 2026-03-06T13:51:29.110 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled container.foo update... 2026-03-06T13:51:29.110 INFO:teuthology.orchestra.run.vm00.stdout:Scheduled container.bar update... 2026-03-06T13:51:29.259 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-06T13:51:29.261 INFO:tasks.cephadm:Waiting for ceph service mon to start (timeout 300)... 2026-03-06T13:51:29.261 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:29.595 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:30.045 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:30.045 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:10.011463Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:10.011542Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:10.011511Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "running": 1, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:30.196 INFO:tasks.cephadm:mon has 2/2 2026-03-06T13:51:30.196 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-06T13:51:30.198 INFO:tasks.cephadm:Waiting for ceph service container.foo to start (timeout 300)... 2026-03-06T13:51:30.198 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:30.243 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: pgmap v35: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:30.243 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:51:30.243 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: Saving service mon spec with placement * 2026-03-06T13:51:30.243 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.243 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: Saving service container.foo spec with placement * 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: Saving service container.bar spec with placement * 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.244 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:30 vm01 ceph-mon[53512]: Deploying daemon container.foo.vm00 on vm00 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: pgmap v35: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='client.14296 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: Saving service mon spec with placement * 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: Saving service container.foo spec with placement * 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: Saving service container.bar spec with placement * 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:30.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:30 vm00 ceph-mon[53145]: Deploying daemon container.foo.vm00 on vm00 2026-03-06T13:51:30.504 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:30.885 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:30.885 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:31.068 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:31.290 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='client.14300 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: Detected new or changed devices on vm01 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: Detected new or changed devices on vm00 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:31 vm01 ceph-mon[53512]: from='client.14304 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='client.14300 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: Detected new or changed devices on vm01 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: Detected new or changed devices on vm00 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:31.388 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:31 vm00 ceph-mon[53145]: from='client.14304 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:32.069 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:32.356 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: pgmap v36: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: pgmap v37: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.668 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:32 vm00 ceph-mon[53145]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:51:32.668 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:32.668 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:31.373504Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:32.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: pgmap v36: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:32.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: pgmap v37: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:32.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:32.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:32 vm01 ceph-mon[53512]: Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-06T13:51:32.815 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:33.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:33 vm00 ceph-mon[53145]: from='client.14308 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:33.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:33 vm01 ceph-mon[53512]: from='client.14308 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:33.816 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:34.103 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:34.408 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:34.408 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:31.373504Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:34.554 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:34.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:34 vm00 ceph-mon[53145]: pgmap v38: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:34.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:34 vm01 ceph-mon[53512]: pgmap v38: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:35.555 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:35.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:35 vm00 ceph-mon[53145]: from='client.14312 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:35.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:35 vm01 ceph-mon[53512]: from='client.14312 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:35.839 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:36.161 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:36.161 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:31.373504Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:36.327 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:36.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:36 vm00 ceph-mon[53145]: pgmap v39: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:36.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:36 vm01 ceph-mon[53512]: pgmap v39: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:37.328 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:37.622 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:37.643 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:37 vm00 ceph-mon[53145]: from='client.14316 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:37.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:37 vm01 ceph-mon[53512]: from='client.14316 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:37.923 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:37.923 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:31.373504Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:38.087 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:38.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:38 vm00 ceph-mon[53145]: pgmap v40: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:38.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:38 vm00 ceph-mon[53145]: from='client.14320 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:38.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:38 vm01 ceph-mon[53512]: pgmap v40: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:38.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:38 vm01 ceph-mon[53512]: from='client.14320 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:39.088 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:39.415 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:39.772 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:39.772 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:29.103023Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:31.373504Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:39.953 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:40.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:40 vm00 ceph-mon[53145]: pgmap v41: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:40.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:40 vm00 ceph-mon[53145]: from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:40.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:40 vm01 ceph-mon[53512]: pgmap v41: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:40.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:40 vm01 ceph-mon[53512]: from='client.14324 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:40.954 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:41.281 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:41.625 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:41.625 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:40.689590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:41.773 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:41.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:41 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:41.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:41 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:41.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:41 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:41.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:41 vm00.local ceph-mon[53145]: Deploying daemon container.foo.vm01 on vm01 2026-03-06T13:51:42.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:41 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:42.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:41 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:42.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:41 vm01 ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:42.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:41 vm01 ceph-mon[53512]: Deploying daemon container.foo.vm01 on vm01 2026-03-06T13:51:42.773 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:42.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:42 vm00.local ceph-mon[53145]: pgmap v42: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:42.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:42 vm00.local ceph-mon[53145]: from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:43.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:42 vm01 ceph-mon[53512]: pgmap v42: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:43.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:42 vm01 ceph-mon[53512]: from='client.14328 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:43.078 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:43.395 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:43.395 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:40.689590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:43.564 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:44.564 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:44.862 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:44.885 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:44 vm00.local ceph-mon[53145]: pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:44.885 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:44 vm00.local ceph-mon[53145]: from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:45.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:44 vm01 ceph-mon[53512]: pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:45.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:44 vm01 ceph-mon[53512]: from='client.14332 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:45.173 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:45.173 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:40.689590Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 0, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:45.322 INFO:tasks.cephadm:container.foo has 0/2 2026-03-06T13:51:46.322 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:46.652 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:46.652 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: pgmap v44: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:46.652 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.653 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.653 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.653 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.653 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:46 vm00.local ceph-mon[53145]: Deploying daemon container.bar.vm01 on vm01 2026-03-06T13:51:46.755 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: from='client.14336 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: pgmap v44: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:46.756 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:46 vm01.local ceph-mon[53512]: Deploying daemon container.bar.vm01 on vm01 2026-03-06T13:51:46.841 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:47.311 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:47.311 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:30.687264Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "running": 1, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:30.687345Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:30.687313Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:30.687374Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:47.481 INFO:tasks.cephadm:container.foo has 1/2 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: pgmap v45: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:48 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.482 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: pgmap v45: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='client.14340 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:48 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:48.787 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:49.103 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:49.103 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.379881Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:47.380053Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.380029Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.379961Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.379929Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.379999Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:49.399 INFO:tasks.cephadm:container.foo has 2/2 2026-03-06T13:51:49.399 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-06T13:51:49.401 INFO:tasks.cephadm:Waiting for ceph service container.bar to start (timeout 300)... 2026-03-06T13:51:49.401 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:49.703 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:49.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:49 vm00.local ceph-mon[53145]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:51:49.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:49 vm00.local ceph-mon[53145]: Cluster is now healthy 2026-03-06T13:51:49.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:49 vm01.local ceph-mon[53512]: Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-06T13:51:49.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:49 vm01.local ceph-mon[53512]: Cluster is now healthy 2026-03-06T13:51:50.020 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:50.021 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.379881Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:47.380053Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.380029Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.379961Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.379929Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.379999Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:50.188 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:51:50.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:50 vm00.local ceph-mon[53145]: from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:50.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:50 vm00.local ceph-mon[53145]: pgmap v46: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:50.389 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:50 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:50.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:50 vm01.local ceph-mon[53512]: from='client.14344 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:50.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:50 vm01.local ceph-mon[53512]: pgmap v46: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:50.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:50 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:51.189 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:51.486 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:51.523 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:51 vm00.local ceph-mon[53145]: from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:51.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:51 vm01.local ceph-mon[53512]: from='client.14348 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:51.857 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:51.857 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.379881Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:29.108449Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:47.380053Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.380029Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.379961Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.379929Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.379999Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:52.040 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:51:52.540 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:52 vm01.local ceph-mon[53512]: pgmap v47: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:52.540 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:52 vm01.local ceph-mon[53512]: from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:52.878 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:52 vm00.local ceph-mon[53145]: pgmap v47: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:52.879 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:52 vm00.local ceph-mon[53145]: from='client.14352 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:53.041 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:53.456 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: Deploying daemon container.bar.vm00 on vm00 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:53 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:53.791 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.470961Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:52.545121Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:53.465076Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.471032Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.470939Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.470897Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.471002Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: Deploying daemon container.bar.vm00 on vm00 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:53 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:51:53.937 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:51:54.938 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:54.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:54 vm00.local ceph-mon[53145]: pgmap v48: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:54.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:54 vm00.local ceph-mon[53145]: from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:55.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:54 vm01.local ceph-mon[53512]: pgmap v48: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:55.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:54 vm01.local ceph-mon[53512]: from='client.14356 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:55.303 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:55.648 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:55.649 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.470961Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:52.545121Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:53.465076Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.471032Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.470939Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.470897Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.471002Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:55.822 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:51:56.824 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:56.852 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:56 vm00.local ceph-mon[53145]: pgmap v49: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:56.852 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:56 vm00.local ceph-mon[53145]: from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:57.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:56 vm01.local ceph-mon[53512]: pgmap v49: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:57.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:56 vm01.local ceph-mon[53512]: from='client.14360 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:57.177 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:57.630 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:57.630 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.470961Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:52.545121Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:53.465076Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.471032Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.470939Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.470897Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.471002Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:57.783 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:51:58.784 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:51:58.841 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:58 vm00.local ceph-mon[53145]: pgmap v50: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:58.842 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:51:58 vm00.local ceph-mon[53145]: from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:59.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:58 vm01.local ceph-mon[53512]: pgmap v50: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:51:59.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:51:58 vm01.local ceph-mon[53512]: from='client.14364 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:51:59.445 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:51:59.812 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:51:59.813 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.470961Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:52.545121Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:51:53.465076Z", "running": 0, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.471032Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.470939Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:51:47.470897Z", "running": 2, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.471002Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:51:59.979 INFO:tasks.cephadm:container.bar has 0/2 2026-03-06T13:52:00.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: pgmap v51: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:00.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:52:00.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:00.710 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-mon[53145]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:00.979 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: pgmap v51: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='client.14368 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:00.986 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:00 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:01.011 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local systemd[1]: Stopping Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:01.011 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[53120]: 2026-03-06T12:52:00.949+0000 7f0687959640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:52:01.011 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:00 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[53120]: 2026-03-06T12:52:00.949+0000 7f0687959640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-06T13:52:01.265 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70151]: 2026-03-06 13:52:01.041574729 +0100 CET m=+0.120104508 container died 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:52:01.265 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70151]: 2026-03-06 13:52:01.177691438 +0100 CET m=+0.256221217 container remove 73e56be0b27efb8971e0f6734be279ef18e0e31ec0e34c645f210e60f2501ee1 (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:52:01.265 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local bash[70151]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a 2026-03-06T13:52:01.414 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:52:01.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service: Deactivated successfully. 2026-03-06T13:52:01.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local systemd[1]: Stopped Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:01.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service: Consumed 2.002s CPU time. 2026-03-06T13:52:01.530 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local systemd[1]: Starting Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:01.781 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70478]: 2026-03-06 13:52:01.689870086 +0100 CET m=+0.024547332 container create 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:52:01.943 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:52:01.943 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:51:47.470961Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:59.985599Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:52:01.865798Z", "running": 1, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:51:47.471032Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:51:47.470939Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:52:01.865666Z", "running": 1, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:51:47.471002Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70478]: 2026-03-06 13:52:01.681752631 +0100 CET m=+0.016429887 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70478]: 2026-03-06 13:52:01.793719544 +0100 CET m=+0.128396799 container init 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local podman[70478]: 2026-03-06 13:52:01.805753675 +0100 CET m=+0.140430921 container start 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local bash[70478]: 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local systemd[1]: Started Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 6 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: pidfile_write: ignore empty --pid-file 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: load: jerasure load: lrc 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Git sha 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: DB SUMMARY 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: DB Session ID: 8CV6IV5C8PWU0JDCXVUN 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: MANIFEST file: MANIFEST-000015 size: 281 Bytes 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000014.log size: 4356989 ; 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.env: 0x56272af36ca0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.info_log: 0x56272bbeab80 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.statistics: (nil) 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.use_fsync: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.db_log_dir: 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.wal_dir: 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.write_buffer_manager: 0x56272bbef900 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:52:02.077 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.unordered_write: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.row_cache: None 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.wal_filter: None 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.wal_compression: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_open_files: -1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Compression algorithms supported: 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kZSTD supported: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.merge_operator: 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_filter: None 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x56272bbea000) 2026-03-06T13:52:02.078 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: index_type: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_index_type: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: index_shortening: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: checksum: 4 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: no_block_cache: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache: 0x56272bc0f1f0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_options: 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: capacity : 536870912 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: num_shard_bits : 4 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: strict_capacity_limit : 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_compressed: (nil) 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: persistent_cache: (nil) 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size: 4096 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size_deviation: 10 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_restart_interval: 16 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: index_block_restart_interval: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: metadata_block_size: 4096 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: partition_filters: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: use_delta_encoding: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: filter_policy: bloomfilter 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: whole_key_filtering: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: verify_compression: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: format_version: 5 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_index_compression: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: block_align: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: max_auto_readahead_size: 262144 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: prepopulate_block_cache: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression: NoCompression 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.num_levels: 7 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:52:02.079 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.ttl: 2592000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:52:02.080 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 245, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 757bbabe-44c3-4b47-8da1-607c9041035c 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801521897837, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801521925880, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 3767898, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 250, "largest_seqno": 3272, "table_properties": {"data_size": 3755915, "index_size": 7480, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 3461, "raw_key_size": 31186, "raw_average_key_size": 22, "raw_value_size": 3728580, "raw_average_value_size": 2703, "num_data_blocks": 356, "num_entries": 1379, "num_filter_entries": 1379, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801521, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "757bbabe-44c3-4b47-8da1-607c9041035c", "db_session_id": "8CV6IV5C8PWU0JDCXVUN", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801521926024, "job": 1, "event": "recovery_finished"} 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x56272bc10e00 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: DB pointer 0x56272bd1c000 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: ** DB Stats ** 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: L0 3/0 3.68 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 164.7 0.02 0.00 1 0.022 0 0 0.0 0.0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Sum 3/0 3.68 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 164.7 0.02 0.00 1 0.022 0 0 0.0 0.0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 164.7 0.02 0.00 1 0.022 0 0 0.0 0.0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 164.7 0.02 0.00 1 0.022 0 0 0.0 0.0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Flush(GB): cumulative 0.004, interval 0.004 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative compaction: 0.00 GB write, 107.78 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval compaction: 0.00 GB write, 107.78 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache BinnedLRUCache@0x56272bc0f1f0#6 capacity: 512.00 MB usage: 42.47 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 5e-06 secs_since: 0 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache entry stats(count,size,portion): DataBlock(4,29.19 KB,0.00556707%) FilterBlock(3,4.80 KB,0.000914931%) IndexBlock(3,8.48 KB,0.00161827%) Misc(1,0.00 KB,0%) 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: starting mon.a rank 0 at public addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] at bind addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???) e2 preinit fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).mds e1 new map 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).mds e1 print_map 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: e1 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: btime 2026-03-06T12:49:40:762524+0000 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: legacy client fscid: -1 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout: No filesystems configured 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).osd e13 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:02.081 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:01 vm00.local ceph-mon[70531]: mon.a@-1(???).paxosservice(auth 1..9) refresh upgraded, format 0 -> 3 2026-03-06T13:52:02.112 INFO:tasks.cephadm:container.bar has 1/2 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: Redeploying mon.a, (container cli args changed) . . . 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: Deploying daemon mon.a on vm00 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: pgmap v52: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='client.24205 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: mon.a calling monitor election 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: monmap epoch 2 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: min_mon_release 19 (squid) 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: election_strategy: 1 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: fsmap 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: mgrmap e14: a(active, since 97s), standbys: b 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: overall HEALTH_OK 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.911 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:02 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: Redeploying mon.a, (container cli args changed) . . . 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: Deploying daemon mon.a on vm00 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: pgmap v52: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='client.24205 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: mon.a calling monitor election 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: monmap epoch 2 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: min_mon_release 19 (squid) 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: election_strategy: 1 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: fsmap 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: mgrmap e14: a(active, since 97s), standbys: b 2026-03-06T13:52:02.943 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: overall HEALTH_OK 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:02.944 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:02 vm01.local ceph-mon[53512]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:03.113 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 -- ceph orch ls -f json 2026-03-06T13:52:03.433 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:52:03.513 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: Stopping Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:03.513 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[53469]: 2026-03-06T12:52:03.342+0000 7f78ca74d640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:52:03.513 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[53469]: 2026-03-06T12:52:03.342+0000 7f78ca74d640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61631]: 2026-03-06 13:52:03.511439294 +0100 CET m=+0.183056839 container died 53426e1156a0262843e0548f9754226c3a6354f180473e3023a14d82bd87d99b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61631]: 2026-03-06 13:52:03.634555673 +0100 CET m=+0.306173208 container remove 53426e1156a0262843e0548f9754226c3a6354f180473e3023a14d82bd87d99b (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local bash[61631]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b.service: Deactivated successfully. 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: Stopped Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:03.764 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b.service: Consumed 1.351s CPU time. 2026-03-06T13:52:03.770 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:52:03.770 INFO:teuthology.orchestra.run.vm00.stdout:[{"events": ["2026-03-06T12:50:35.900250Z service:agent [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "agent", "service_type": "agent", "status": {"created": "2026-03-06T12:50:25.015507Z", "last_refresh": "2026-03-06T12:52:02.709521Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "set -e\ntest -f /var/cache/bar/from.txt\ntest -f /var/cache/bar/presized.dat\necho ok > /var/cache/bar/primary.txt\nsleep infinity\n", "mount_path": "/root/init_check.sh"}], "events": ["2026-03-06T12:51:59.985599Z service:container.bar [INFO] \"service was created\""], "extra_entrypoint_args": ["/root/init_check.sh"], "placement": {"host_pattern": "*"}, "service_id": "bar", "service_name": "container.bar", "service_type": "container", "spec": {"dirs": ["data"], "entrypoint": "bash", "image": "quay.io/fedora/fedora:latest", "init_containers": [{"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": ". /etc/os-release && echo from=$ID > /var/cache/bar/from.txt", "split": false}], "image": "quay.io/centos/centos:latest", "volume_mounts": {"data": "/var/cache/bar:z"}}, {"entrypoint": "bash", "entrypoint_args": [{"argument": "-c", "split": false}, {"argument": "test -f /var/cache/bar/from.txt && truncate -s 102400 /var/cache/bar/presized.dat", "split": false}], "volume_mounts": {"data": "/var/cache/bar:z"}}], "volume_mounts": {"data": "/var/cache/bar:z"}}, "status": {"created": "2026-03-06T12:51:29.103930Z", "last_refresh": "2026-03-06T12:52:02.709703Z", "running": 2, "size": 2}}, {"custom_configs": [{"content": "while getopts \"o:c:\" opt; do\n case ${opt} in\n o )\n OUT_FILE=${OPTARG}\n ;;\n c )\n CONTENT=${OPTARG}\n esac\ndone\necho $CONTENT > $OUT_FILE\nsleep infinity\n", "mount_path": "/root/write_thing_to_file.sh"}], "events": ["2026-03-06T12:51:45.496356Z service:container.foo [INFO] \"service was created\""], "extra_container_args": ["-v", "/etc/cephadm_testing:/root/cephadm_testing"], "extra_entrypoint_args": ["/root/write_thing_to_file.sh", "-c", "testing_custom_containers", "-o", "/root/cephadm_testing/testing.txt"], "placement": {"host_pattern": "*"}, "service_id": "foo", "service_name": "container.foo", "service_type": "container", "spec": {"entrypoint": "bash", "image": "quay.io/fedora/fedora:latest"}, "status": {"created": "2026-03-06T12:51:29.099948Z", "last_refresh": "2026-03-06T12:52:02.709657Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:50:48.083070Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm00=a", "vm01=b"]}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-06T12:50:46.979788Z", "last_refresh": "2026-03-06T12:52:02.709600Z", "running": 2, "size": 2}}, {"events": ["2026-03-06T12:51:29.113142Z service:mon [INFO] \"service was created\""], "extra_container_args": ["--cpus=2"], "extra_entrypoint_args": ["--debug_ms 10"], "placement": {"host_pattern": "*"}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-06T12:51:29.095705Z", "last_refresh": "2026-03-06T12:52:02.709569Z", "running": 1, "size": 2}}, {"service_name": "osd", "service_type": "osd", "spec": {"filter_logic": "AND", "objectstore": "bluestore"}, "status": {"container_image_id": "306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68", "container_image_name": "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b", "last_refresh": "2026-03-06T12:52:02.709629Z", "running": 2, "size": 2}, "unmanaged": true}] 2026-03-06T13:52:03.930 INFO:tasks.cephadm:container.bar has 2/2 2026-03-06T13:52:03.930 INFO:teuthology.run_tasks:Running task exec... 2026-03-06T13:52:03.932 INFO:teuthology.task.exec:Executing custom commands... 2026-03-06T13:52:03.932 INFO:teuthology.task.exec:Running commands on role host.a host ubuntu@vm00.local 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> sudo TESTDIR=/home/ubuntu/cephtest bash -c 'set -ex 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> FSID=$(/home/ubuntu/cephtest/cephadm shell -- ceph fsid) 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> sleep 60 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # check extra container and entrypoint args written to mon unit run file 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> grep "\-\-cpus=2" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> grep "\-\-debug_ms 10" /var/lib/ceph/$FSID/mon.*/unit.run 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # check that custom container properly wrote content to file. 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # This requires the custom config, extra container args, and 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # entrypoint args to all be working in order for this to have 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # been written. The container entrypoint was set up with custom_configs, 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # the content and where to write to with the entrypoint args, and the mounting 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # of the /etc/cephadm_testing dir with extra container args 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> grep "testing_custom_containers" /etc/cephadm_testing/testing.txt 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # Verify that container bar'"'"'s init containers and primary container 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> # ran successfully 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> dir=$(find /var/lib/ceph/$FSID -maxdepth 1 -type d -name '"'"'container.bar.*'"'"') 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> test -n "$dir" 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> grep ok ${dir}/data/primary.txt 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> grep from=centos ${dir}/data/from.txt 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> test -s ${dir}/data/presized.dat 2026-03-06T13:52:03.932 DEBUG:teuthology.orchestra.run.vm00:> ' 2026-03-06T13:52:03.958 INFO:teuthology.orchestra.run.vm00.stderr:++ /home/ubuntu/cephtest/cephadm shell -- ceph fsid 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: Starting Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61751]: 2026-03-06 13:52:03.945436192 +0100 CET m=+0.016644938 container create fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61751]: 2026-03-06 13:52:03.979252431 +0100 CET m=+0.050461177 container init fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61751]: 2026-03-06 13:52:03.986483396 +0100 CET m=+0.057692142 container start fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local bash[61751]: fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local podman[61751]: 2026-03-06 13:52:03.93841584 +0100 CET m=+0.009624597 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:03 vm01.local systemd[1]: Started Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 6 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: pidfile_write: ignore empty --pid-file 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: load: jerasure load: lrc 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Git sha 0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: DB SUMMARY 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: DB Session ID: F2VUVX1AG13ERMXFA89Z 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 1, files: 000008.sst 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000009.log size: 5885555 ; 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.env: 0x55f5ad180ca0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.info_log: 0x55f5ae32c1a0 2026-03-06T13:52:04.038 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.statistics: (nil) 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.use_fsync: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.db_log_dir: 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.wal_dir: 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.write_buffer_manager: 0x55f5ae331900 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.unordered_write: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.row_cache: None 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.wal_filter: None 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.wal_compression: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:52:04.039 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_open_files: -1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Compression algorithms supported: 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kZSTD supported: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.merge_operator: 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_filter: None 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55f5ae32c360) 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: index_type: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: data_block_index_type: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: index_shortening: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: checksum: 4 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: no_block_cache: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache: 0x55f5ae3511f0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_options: 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: capacity : 536870912 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: num_shard_bits : 4 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: strict_capacity_limit : 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_compressed: (nil) 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: persistent_cache: (nil) 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_size: 4096 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_size_deviation: 10 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_restart_interval: 16 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: index_block_restart_interval: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: metadata_block_size: 4096 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: partition_filters: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: use_delta_encoding: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: filter_policy: bloomfilter 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: whole_key_filtering: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: verify_compression: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: format_version: 5 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: enable_index_compression: 1 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: block_align: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: max_auto_readahead_size: 262144 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: prepopulate_block_cache: 0 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression: NoCompression 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:52:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.num_levels: 7 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:52:04.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.ttl: 2592000 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7a8e4116-c05f-4095-bcc3-657bdde48fef 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801524015485, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801524032416, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 4068509, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 6, "largest_seqno": 3258, "table_properties": {"data_size": 4055150, "index_size": 8216, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 4101, "raw_key_size": 42156, "raw_average_key_size": 25, "raw_value_size": 4021227, "raw_average_value_size": 2460, "num_data_blocks": 392, "num_entries": 1634, "num_filter_entries": 1634, "num_deletions": 4, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801524, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7a8e4116-c05f-4095-bcc3-657bdde48fef", "db_session_id": "F2VUVX1AG13ERMXFA89Z", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801524032625, "job": 1, "event": "recovery_finished"} 2026-03-06T13:52:04.042 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-06T13:52:04.239 INFO:teuthology.orchestra.run.vm00.stderr:Inferring fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:04.283 INFO:teuthology.orchestra.run.vm00.stderr:Inferring config /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55f5ae352e00 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: DB pointer 0x55f5ae46a000 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: ** DB Stats ** 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: L0 2/0 3.88 MB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 363.2 0.01 0.00 1 0.011 0 0 0.0 0.0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Sum 2/0 3.88 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 363.2 0.01 0.00 1 0.011 0 0 0.0 0.0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 363.2 0.01 0.00 1 0.011 0 0 0.0 0.0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 363.2 0.01 0.00 1 0.011 0 0 0.0 0.0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Flush(GB): cumulative 0.004, interval 0.004 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative compaction: 0.00 GB write, 144.89 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval compaction: 0.00 GB write, 144.89 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache BinnedLRUCache@0x55f5ae3511f0#6 capacity: 512.00 MB usage: 25.12 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 4e-06 secs_since: 0 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache entry stats(count,size,portion): DataBlock(3,11.30 KB,0.00215471%) FilterBlock(2,4.64 KB,0.000885129%) IndexBlock(2,9.19 KB,0.00175238%) Misc(1,0.00 KB,0%) 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:04.291 INFO:journalctl@ceph.mon.b.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:52:04.387 INFO:teuthology.orchestra.run.vm00.stderr:Using ceph image with id '306e97de47e9' and tag 'cobaltcore-storage-v19.2.3-fasttrack-3' created on 2026-03-03 21:58:41 +0000 UTC 2026-03-06T13:52:04.387 INFO:teuthology.orchestra.run.vm00.stderr:harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: purged_snaps scrub ok 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: purged_snaps scrub ok 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: Redeploying mon.b, (container cli args changed) . . . 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: Deploying daemon mon.b on vm01 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: pgmap v53: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:04.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:04 vm00.local ceph-mon[70531]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: purged_snaps scrub ok 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: purged_snaps scrub ok 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: Redeploying mon.b, (container cli args changed) . . . 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: Deploying daemon mon.b on vm01 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: pgmap v53: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:04.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:04 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:04.948 INFO:teuthology.orchestra.run.vm00.stderr:+ FSID=e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:04.948 INFO:teuthology.orchestra.run.vm00.stderr:+ sleep 60 2026-03-06T13:52:05.267 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local systemd[1]: Stopping Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:05.268 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[70489]: 2026-03-06T12:52:05.131+0000 7f77d5ebe640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:52:05.268 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[70489]: 2026-03-06T12:52:05.131+0000 7f77d5ebe640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-06T13:52:05.268 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71316]: 2026-03-06 13:52:05.154396773 +0100 CET m=+0.036013230 container died 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:52:05.551 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71316]: 2026-03-06 13:52:05.269954482 +0100 CET m=+0.151570949 container remove 84cbe49b28f85f57bd4f706f8027313df263da90f2eb5ce604a4eb311ca36932 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:52:05.551 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local bash[71316]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a 2026-03-06T13:52:05.551 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service: Deactivated successfully. 2026-03-06T13:52:05.551 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local systemd[1]: Stopped Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:05.551 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local systemd[1]: Starting Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71434]: 2026-03-06 13:52:05.549548249 +0100 CET m=+0.013786145 container create 0edab4b746755a89fd50dba84ce83e2c1b7a081b5493524519b348bf2269ffbb (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71434]: 2026-03-06 13:52:05.583695832 +0100 CET m=+0.047933728 container init 0edab4b746755a89fd50dba84ce83e2c1b7a081b5493524519b348bf2269ffbb (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71434]: 2026-03-06 13:52:05.587384051 +0100 CET m=+0.051621937 container start 0edab4b746755a89fd50dba84ce83e2c1b7a081b5493524519b348bf2269ffbb (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local bash[71434]: 0edab4b746755a89fd50dba84ce83e2c1b7a081b5493524519b348bf2269ffbb 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local podman[71434]: 2026-03-06 13:52:05.543953168 +0100 CET m=+0.008191064 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local systemd[1]: Started Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 7 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: pidfile_write: ignore empty --pid-file 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: load: jerasure load: lrc 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Git sha 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: DB SUMMARY 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: DB Session ID: WIG0UY7NKYGI7W736NDF 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: MANIFEST file: MANIFEST-000020 size: 373 Bytes 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 3, files: 000008.sst 000013.sst 000018.sst 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000019.log size: 360222 ; 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.env: 0x55684d6aaca0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.info_log: 0x55684f0401a0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.statistics: (nil) 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.use_fsync: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.db_log_dir: 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.wal_dir: 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:52:05.959 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.write_buffer_manager: 0x55684f045900 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.unordered_write: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.row_cache: None 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.wal_filter: None 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.wal_compression: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_open_files: -1 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:52:05.960 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Compression algorithms supported: 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kZSTD supported: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.merge_operator: 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_filter: None 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55684f040360) 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: index_type: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_index_type: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: index_shortening: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: checksum: 4 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: no_block_cache: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache: 0x55684f0651f0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_options: 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: capacity : 536870912 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: num_shard_bits : 4 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: strict_capacity_limit : 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_cache_compressed: (nil) 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: persistent_cache: (nil) 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size: 4096 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_size_deviation: 10 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_restart_interval: 16 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: index_block_restart_interval: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: metadata_block_size: 4096 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: partition_filters: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: use_delta_encoding: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: filter_policy: bloomfilter 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: whole_key_filtering: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: verify_compression: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: format_version: 5 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_index_compression: 1 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: block_align: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: max_auto_readahead_size: 262144 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: prepopulate_block_cache: 0 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:52:05.961 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression: NoCompression 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.num_levels: 7 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:52:05.962 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.ttl: 2592000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000020 succeeded,manifest_file_number is 20, next_file_number is 22, last_sequence is 3272, log_number is 15,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 15 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 15 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 757bbabe-44c3-4b47-8da1-607c9041035c 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801525612278, "job": 1, "event": "recovery_started", "wal_files": [19]} 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #19 mode 2 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801525614742, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 23, "file_size": 302695, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3278, "largest_seqno": 3481, "table_properties": {"data_size": 300389, "index_size": 947, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 325, "raw_key_size": 2602, "raw_average_key_size": 24, "raw_value_size": 297691, "raw_average_value_size": 2782, "num_data_blocks": 42, "num_entries": 107, "num_filter_entries": 107, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801525, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "757bbabe-44c3-4b47-8da1-607c9041035c", "db_session_id": "WIG0UY7NKYGI7W736NDF", "orig_file_number": 23, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801525614791, "job": 1, "event": "recovery_finished"} 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/version_set.cc:5047] Creating manifest 25 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000019.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55684f066e00 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: DB pointer 0x55684f182000 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/compaction/compaction_job.cc:1995] [default] [JOB 3] Compacting 4@0 files to L6, score 1.00 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/compaction/compaction_job.cc:2001] [default]: Compaction start summary: Base version 2 Base level 0, inputs: [23(295KB) 18(3679KB) 13(82KB) 8(1900B)] 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801525616835, "job": 3, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [23, 18, 13, 8], "score": 1, "input_data_size": 4156531, "oldest_snapshot_seqno": -1} 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: CompressorRegistry(0x55684f02d418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: ** DB Stats ** 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: L0 4/4 3.96 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 141.3 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Sum 4/4 3.96 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 141.3 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 141.3 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 141.3 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-06T13:52:05.963 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Cumulative compaction: 0.00 GB write, 26.26 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Interval compaction: 0.00 GB write, 26.26 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache BinnedLRUCache@0x55684f0651f0#7 capacity: 512.00 MB usage: 60.73 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 5e-06 secs_since: 0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: Block cache entry stats(count,size,portion): DataBlock(5,33.77 KB,0.00644028%) FilterBlock(4,5.17 KB,0.000986457%) IndexBlock(4,9.56 KB,0.0018239%) Misc(3,8.16 KB,0.00155568%) 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: Event(0x556850094088 nevent=5000 time_id=1).set_owner center_id=1 owner=140108530026048 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: stack operator() starting 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: Event(0x556850094508 nevent=5000 time_id=1).set_owner center_id=2 owner=140108622345792 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: stack operator() starting 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: Event(0x55684f05fdc8 nevent=5000 time_id=1).set_owner center_id=0 owner=140108521633344 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: stack operator() starting 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: starting mon.a rank 0 at public addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] at bind addrs [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: CompressorRegistry(0x55684f02dd18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???) e2 preinit fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).mds e1 new map 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).mds e1 print_map 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: e1 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: btime 2026-03-06T12:49:40:762524+0000 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: legacy client fscid: -1 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout: No filesystems configured 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).osd e13 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:05.964 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:05 vm00.local ceph-mon[71470]: mon.a@-1(???).paxosservice(auth 1..10) refresh upgraded, format 0 -> 3 2026-03-06T13:52:06.540 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: pgmap v54: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: mon.a calling monitor election 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: monmap epoch 2 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: min_mon_release 19 (squid) 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: election_strategy: 1 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: fsmap 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: mgrmap e14: a(active, since 101s), standbys: b 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: overall HEALTH_OK 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='' 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:06.541 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-mon[61784]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:06.708 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: pgmap v54: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: mon.a calling monitor election 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: monmap epoch 2 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: min_mon_release 19 (squid) 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: election_strategy: 1 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: fsmap 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: mgrmap e14: a(active, since 101s), standbys: b 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: overall HEALTH_OK 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 ' entity='' 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 ' entity='mgr.a' 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-06T13:52:06.709 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:06 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:07.150 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local systemd[1]: Stopping Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:07.150 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[61761]: 2026-03-06T12:52:06.951+0000 7ffbba4c8640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:52:07.150 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:06 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[61761]: 2026-03-06T12:52:06.951+0000 7ffbba4c8640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-06T13:52:07.150 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62013]: 2026-03-06 13:52:07.036680895 +0100 CET m=+0.098207712 container died fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:52:07.435 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62013]: 2026-03-06 13:52:07.153312894 +0100 CET m=+0.214839711 container remove fb9a7fe813e037b6956e74c7bf1f666456504c4fad5dc9069d908e76d85808d0 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:52:07.435 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local bash[62013]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b 2026-03-06T13:52:07.435 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b.service: Deactivated successfully. 2026-03-06T13:52:07.435 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local systemd[1]: Stopped Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:07.435 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local systemd[1]: Starting Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62130]: 2026-03-06 13:52:07.433725792 +0100 CET m=+0.015433451 container create e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62130]: 2026-03-06 13:52:07.462467293 +0100 CET m=+0.044174952 container init e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62130]: 2026-03-06 13:52:07.466188038 +0100 CET m=+0.047895687 container start e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True) 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local bash[62130]: e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local podman[62130]: 2026-03-06 13:52:07.427776576 +0100 CET m=+0.009484236 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local systemd[1]: Started Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2. 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: set uid:gid to 167:167 (ceph:ceph) 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: ceph version 19.2.3-47-gc24117fd552 (c24117fd5525679b799527bc1bd1f1dd0a2db5e2) squid (stable), process ceph-mon, pid 6 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: pidfile_write: ignore empty --pid-file 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: load: jerasure load: lrc 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: RocksDB version: 7.9.2 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Git sha 0 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Compile date 2026-03-03 21:08:28 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: DB SUMMARY 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: DB Session ID: EO7GWTYTW2PUYDFNOY9E 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: CURRENT file: CURRENT 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: IDENTITY file: IDENTITY 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: MANIFEST file: MANIFEST-000015 size: 282 Bytes 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: SST files in /var/lib/ceph/mon/ceph-b/store.db dir, Total Num: 2, files: 000008.sst 000013.sst 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-b/store.db: 000014.log size: 195470 ; 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.error_if_exists: 0 2026-03-06T13:52:07.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.create_if_missing: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.paranoid_checks: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.env: 0x55d89b74cca0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.fs: PosixFileSystem 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.info_log: 0x55d89d2381a0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_file_opening_threads: 16 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.statistics: (nil) 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.use_fsync: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_log_file_size: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.keep_log_file_num: 1000 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.recycle_log_file_num: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_fallocate: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_mmap_reads: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_mmap_writes: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.use_direct_reads: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.create_missing_column_families: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.db_log_dir: 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.wal_dir: 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.advise_random_on_open: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.db_write_buffer_size: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.write_buffer_manager: 0x55d89d23d900 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.rate_limiter: (nil) 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.wal_recovery_mode: 2 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enable_thread_tracking: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enable_pipelined_write: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.unordered_write: 0 2026-03-06T13:52:07.792 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.row_cache: None 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.wal_filter: None 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_ingest_behind: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.two_write_queues: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.manual_wal_flush: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.wal_compression: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.atomic_flush: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.log_readahead_size: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.best_efforts_recovery: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.allow_data_in_errors: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.db_host_id: __hostname__ 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_background_jobs: 2 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_background_compactions: -1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_subcompactions: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_total_wal_size: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_open_files: -1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bytes_per_sync: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_readahead_size: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_background_flushes: -1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Compression algorithms supported: 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kZSTD supported: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kXpressCompression supported: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kBZip2Compression supported: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kLZ4Compression supported: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kZlibCompression supported: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kLZ4HCCompression supported: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: kSnappyCompression supported: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.merge_operator: 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_filter: None 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_filter_factory: None 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.sst_partitioner_factory: None 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55d89d238360) 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: cache_index_and_filter_blocks: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: pin_top_level_index_and_filter: 1 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: index_type: 0 2026-03-06T13:52:07.793 INFO:journalctl@ceph.mon.b.vm01.stdout: data_block_index_type: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: index_shortening: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: checksum: 4 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: no_block_cache: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache: 0x55d89d25d1f0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_name: BinnedLRUCache 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_options: 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: capacity : 536870912 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: num_shard_bits : 4 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: strict_capacity_limit : 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: high_pri_pool_ratio: 0.000 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_cache_compressed: (nil) 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: persistent_cache: (nil) 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_size: 4096 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_size_deviation: 10 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_restart_interval: 16 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: index_block_restart_interval: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: metadata_block_size: 4096 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: partition_filters: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: use_delta_encoding: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: filter_policy: bloomfilter 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: whole_key_filtering: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: verify_compression: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: read_amp_bytes_per_bit: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: format_version: 5 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: enable_index_compression: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: block_align: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: max_auto_readahead_size: 262144 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: prepopulate_block_cache: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: initial_auto_readahead_size: 8192 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout: num_file_reads_for_auto_readahead: 2 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.write_buffer_size: 33554432 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_write_buffer_number: 2 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression: NoCompression 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression: Disabled 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.prefix_extractor: nullptr 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.num_levels: 7 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.level: 32767 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.strategy: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.enabled: false 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.target_file_size_base: 67108864 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-06T13:52:07.794 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.arena_block_size: 1048576 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.disable_auto_compactions: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.inplace_update_support: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.bloom_locality: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.max_successive_merges: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.paranoid_file_checks: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.force_consistency_checks: 1 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.report_bg_io_stats: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.ttl: 2592000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enable_blob_files: false 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.min_blob_size: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_file_size: 268435456 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.blob_file_starting_level: 0 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-b/store.db/MANIFEST-000015 succeeded,manifest_file_number is 15, next_file_number is 17, last_sequence is 3258, log_number is 10,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 10 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 10 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 7a8e4116-c05f-4095-bcc3-657bdde48fef 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801527488259, "job": 1, "event": "recovery_started", "wal_files": [14]} 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #14 mode 2 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801527489895, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 18, "file_size": 128329, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 3270, "largest_seqno": 3378, "table_properties": {"data_size": 126635, "index_size": 463, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 197, "raw_key_size": 1897, "raw_average_key_size": 26, "raw_value_size": 124901, "raw_average_value_size": 1734, "num_data_blocks": 18, "num_entries": 72, "num_filter_entries": 72, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772801527, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "7a8e4116-c05f-4095-bcc3-657bdde48fef", "db_session_id": "EO7GWTYTW2PUYDFNOY9E", "orig_file_number": 18, "seqno_to_time_mapping": "N/A"}} 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: EVENT_LOG_v1 {"time_micros": 1772801527489943, "job": 1, "event": "recovery_finished"} 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/version_set.cc:5047] Creating manifest 20 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-b/store.db/000014.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55d89d25ee00 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: DB pointer 0x55d89d37a000 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-06T13:52:07.795 INFO:journalctl@ceph.mon.b.vm01.stdout: ** DB Stats ** 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: L0 3/0 4.00 MB 0.8 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 87.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Sum 3/0 4.00 MB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 87.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 87.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: ** Compaction Stats [default] ** 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 87.2 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Uptime(secs): 0.0 total, 0.0 interval 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Flush(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Total Files): cumulative 0, interval 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(L0 Files): cumulative 0, interval 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: AddFile(Keys): cumulative 0, interval 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Cumulative compaction: 0.00 GB write, 20.39 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Interval compaction: 0.00 GB write, 20.39 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache BinnedLRUCache@0x55d89d25d1f0#6 capacity: 512.00 MB usage: 14.61 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 5e-06 secs_since: 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: Block cache entry stats(count,size,portion): FilterBlock(3,4.88 KB,0.000929832%) IndexBlock(3,9.73 KB,0.00185668%) Misc(1,0.00 KB,0%) 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: ** File Read Latency Histogram By Level [default] ** 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: CompressorRegistry(0x55d89d225418) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: Event(0x55d89e296088 nevent=5000 time_id=1).set_owner center_id=0 owner=140509116450368 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: stack operator() starting 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: Event(0x55d89e2962c8 nevent=5000 time_id=1).set_owner center_id=1 owner=140509124843072 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: stack operator() starting 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: Event(0x55d89e296508 nevent=5000 time_id=1).set_owner center_id=2 owner=140509217162816 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: stack operator() starting 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: starting mon.b rank 1 at public addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] at bind addrs [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon_data /var/lib/ceph/mon/ceph-b fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: CompressorRegistry(0x55d89d225d18) _refresh_config ms_osd_compression_mode 0 ms_osd_compression_methods [1] ms_osd_compress_above_min_size 1024 ms_compress_secure 0 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???) e2 preinit fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mds e0 Unable to load 'last_metadata' 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mds e0 Unable to load 'last_metadata' 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mds e1 new map 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mds e1 print_map 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: e1 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: btime 2026-03-06T12:49:40:762524+0000 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: legacy client fscid: -1 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout: No filesystems configured 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).osd e13 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).osd e13 crush map has features 288514050185494528, adjusting msgr requires 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).paxosservice(auth 1..11) refresh upgraded, format 0 -> 3 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mgr e0 loading version 14 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mgr e14 active server: [v2:192.168.123.100:6800/1074681661,v1:192.168.123.100:6801/1074681661](14156) 2026-03-06T13:52:07.796 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: mon.b@-1(???).mgr e14 mkfs or daemon transitioned to available, loading commands 2026-03-06T13:52:08.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:07 vm00.local ceph-mon[71470]: Redeploying mon.b, (entrypoint args changed) . . . 2026-03-06T13:52:08.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:07 vm00.local ceph-mon[71470]: Deploying daemon mon.b on vm01 2026-03-06T13:52:08.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:07 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:08.209 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:07 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:08.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: Redeploying mon.b, (entrypoint args changed) . . . 2026-03-06T13:52:08.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: Deploying daemon mon.b on vm01 2026-03-06T13:52:08.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-06T13:52:08.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:07 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: pgmap v55: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: mon.b calling monitor election 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: mon.a calling monitor election 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: monmap epoch 2 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: min_mon_release 19 (squid) 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: election_strategy: 1 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: fsmap 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: mgrmap e14: a(active, since 102s), standbys: b 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: overall HEALTH_OK 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:52:09.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:08 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: pgmap v55: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: mon.b calling monitor election 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: mon.a calling monitor election 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: monmap epoch 2 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: last_changed 2026-03-06T12:50:40.008473+0000 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: created 2026-03-06T12:49:39.578527+0000 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: min_mon_release 19 (squid) 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: election_strategy: 1 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: 0: [v2:192.168.123.100:3300/0,v1:192.168.123.100:6789/0] mon.a 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: 1: [v2:192.168.123.101:3300/0,v1:192.168.123.101:6789/0] mon.b 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: fsmap 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: osdmap e13: 2 total, 2 up, 2 in 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: mgrmap e14: a(active, since 102s), standbys: b 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: overall HEALTH_OK 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-06T13:52:09.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-06T13:52:09.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-06T13:52:09.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:08 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:11.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:10 vm01.local ceph-mon[62162]: pgmap v56: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:11.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:10 vm00.local ceph-mon[71470]: pgmap v56: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:12.291 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:11 vm01.local ceph-mon[62162]: pgmap v57: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:12.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:11 vm00.local ceph-mon[71470]: pgmap v57: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:14.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:14 vm01.local ceph-mon[62162]: pgmap v58: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:14.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:14 vm00.local ceph-mon[71470]: pgmap v58: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:16.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:16 vm01.local ceph-mon[62162]: pgmap v59: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:16.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:16 vm00.local ceph-mon[71470]: pgmap v59: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:18.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:18 vm01.local ceph-mon[62162]: pgmap v60: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:18.828 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:18 vm00.local ceph-mon[71470]: pgmap v60: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:20.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:20 vm01.local ceph-mon[62162]: pgmap v61: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:20.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:20 vm00.local ceph-mon[71470]: pgmap v61: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: pgmap v62: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.791 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:22 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: pgmap v62: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:22.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:22 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:24.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:24 vm01.local ceph-mon[62162]: pgmap v63: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:24.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:24 vm00.local ceph-mon[71470]: pgmap v63: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:26.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:26 vm01.local ceph-mon[62162]: pgmap v64: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:26.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:26 vm00.local ceph-mon[71470]: pgmap v64: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:28.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:28 vm01.local ceph-mon[62162]: pgmap v65: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:28.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:28 vm00.local ceph-mon[71470]: pgmap v65: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:30.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:30 vm00.local ceph-mon[71470]: pgmap v66: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:31.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:30 vm01.local ceph-mon[62162]: pgmap v66: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:32.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:32 vm00.local ceph-mon[71470]: pgmap v67: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:33.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:32 vm01.local ceph-mon[62162]: pgmap v67: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:34.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:34 vm00.local ceph-mon[71470]: pgmap v68: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:35.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:34 vm01.local ceph-mon[62162]: pgmap v68: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:36.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:36 vm00.local ceph-mon[71470]: pgmap v69: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:37.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:36 vm01.local ceph-mon[62162]: pgmap v69: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:38.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:38 vm01.local ceph-mon[62162]: pgmap v70: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:38.848 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:38 vm00.local ceph-mon[71470]: pgmap v70: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:40.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:40 vm00.local ceph-mon[71470]: pgmap v71: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:41.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:40 vm01.local ceph-mon[62162]: pgmap v71: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:42.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:42 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:42 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:42 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.459 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:42 vm00.local ceph-mon[71470]: pgmap v72: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:42.496 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:42 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.496 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:42 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.496 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:42 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:42.496 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:42 vm01.local ceph-mon[62162]: pgmap v72: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:43 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:43.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:43 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:52:44.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:44 vm00.local ceph-mon[71470]: pgmap v73: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:45.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:44 vm01.local ceph-mon[62162]: pgmap v73: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:46.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:46 vm00.local ceph-mon[71470]: pgmap v74: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:47.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:46 vm01.local ceph-mon[62162]: pgmap v74: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:48.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:48 vm00.local ceph-mon[71470]: pgmap v75: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:49.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:48 vm01.local ceph-mon[62162]: pgmap v75: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:50.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:50 vm00.local ceph-mon[71470]: pgmap v76: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:51.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:50 vm01.local ceph-mon[62162]: pgmap v76: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:52.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:52 vm00.local ceph-mon[71470]: pgmap v77: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:53.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:52 vm01.local ceph-mon[62162]: pgmap v77: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:54.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:54 vm00.local ceph-mon[71470]: pgmap v78: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:55.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:54 vm01.local ceph-mon[62162]: pgmap v78: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:56.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:56 vm00.local ceph-mon[71470]: pgmap v79: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:57.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:56 vm01.local ceph-mon[62162]: pgmap v79: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:58.790 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:52:58 vm01.local ceph-mon[62162]: pgmap v80: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:52:58.895 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:52:58 vm00.local ceph-mon[71470]: pgmap v80: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:00.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:00 vm00.local ceph-mon[71470]: pgmap v81: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:01.041 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:00 vm01.local ceph-mon[62162]: pgmap v81: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:02.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:02 vm00.local ceph-mon[71470]: pgmap v82: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:02.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:02 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:02.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:02 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:02.458 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:02 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:02.459 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:02 vm01.local ceph-mon[62162]: pgmap v82: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:02.459 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:02 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:02.459 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:02 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:02.459 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:02 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:03.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:03 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:03.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:03 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:03.958 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:03 vm00.local ceph-mon[71470]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:03 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:03 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:04.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:03 vm01.local ceph-mon[62162]: from='mgr.14156 192.168.123.100:0/3512874990' entity='mgr.a' 2026-03-06T13:53:04.949 INFO:teuthology.orchestra.run.vm00.stderr:+ grep '\-\-cpus=2' /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/unit.run 2026-03-06T13:53:04.950 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a --pids-limit=-1 -d --log-driver journald --conmon-pidfile /run/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service-pid --cidfile /run/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service-cid --cgroups=split --cpus=2 -e CONTAINER_IMAGE=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b -e NODE_NAME=vm00.local -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2:/var/run/ceph:z -v /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2:/var/log/ceph:z -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false --debug_ms 10 2026-03-06T13:53:04.950 INFO:teuthology.orchestra.run.vm00.stderr:+ grep '\-\-debug_ms 10' /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/unit.run 2026-03-06T13:53:04.951 INFO:teuthology.orchestra.run.vm00.stdout:/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph-mon --privileged --group-add=disk --init --name ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a --pids-limit=-1 -d --log-driver journald --conmon-pidfile /run/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service-pid --cidfile /run/ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service-cid --cgroups=split --cpus=2 -e CONTAINER_IMAGE=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b -e NODE_NAME=vm00.local -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2:/var/run/ceph:z -v /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2:/var/log/ceph:z -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /dev:/dev -v /run/udev:/run/udev -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a:/var/lib/ceph/mon/ceph-a:z -v /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/mon.a/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false --debug_ms 10 2026-03-06T13:53:04.951 INFO:teuthology.orchestra.run.vm00.stderr:+ grep testing_custom_containers /etc/cephadm_testing/testing.txt 2026-03-06T13:53:04.952 INFO:teuthology.orchestra.run.vm00.stdout:testing_custom_containers 2026-03-06T13:53:04.952 INFO:teuthology.orchestra.run.vm00.stderr:++ find /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2 -maxdepth 1 -type d -name 'container.bar.*' 2026-03-06T13:53:04.953 INFO:teuthology.orchestra.run.vm00.stderr:+ dir=/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/container.bar.vm00 2026-03-06T13:53:04.953 INFO:teuthology.orchestra.run.vm00.stderr:+ test -n /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/container.bar.vm00 2026-03-06T13:53:04.953 INFO:teuthology.orchestra.run.vm00.stderr:+ grep ok /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/container.bar.vm00/data/primary.txt 2026-03-06T13:53:04.954 INFO:teuthology.orchestra.run.vm00.stdout:ok 2026-03-06T13:53:04.954 INFO:teuthology.orchestra.run.vm00.stderr:+ grep from=centos /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/container.bar.vm00/data/from.txt 2026-03-06T13:53:04.955 INFO:teuthology.orchestra.run.vm00.stdout:from=centos 2026-03-06T13:53:04.955 INFO:teuthology.orchestra.run.vm00.stderr:+ test -s /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/container.bar.vm00/data/presized.dat 2026-03-06T13:53:04.955 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:04 vm00.local ceph-mon[71470]: pgmap v83: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:04.957 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-06T13:53:04.959 INFO:tasks.cephadm:Teardown begin 2026-03-06T13:53:04.959 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:53:05.020 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:53:05.040 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:04 vm01.local ceph-mon[62162]: pgmap v83: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-06T13:53:05.045 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-06T13:53:05.045 DEBUG:teuthology.orchestra.run.vm00:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:53:05.074 DEBUG:teuthology.orchestra.run.vm01:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-06T13:53:05.099 INFO:tasks.cephadm:Stopping all daemons... 2026-03-06T13:53:05.099 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-06T13:53:05.099 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a 2026-03-06T13:53:05.208 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:05 vm00.local systemd[1]: Stopping Ceph mon.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:05.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:05 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[71444]: 2026-03-06T12:53:05.204+0000 7f6d962a6640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false --debug_ms 10 (PID: 1) UID: 0 2026-03-06T13:53:05.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:05 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a[71444]: 2026-03-06T12:53:05.204+0000 7f6d962a6640 -1 mon.a@0(leader) e2 *** Got Signal Terminated *** 2026-03-06T13:53:05.509 INFO:journalctl@ceph.mon.a.vm00.stdout:Mar 06 13:53:05 vm00.local podman[72496]: 2026-03-06 13:53:05.396333542 +0100 CET m=+0.207822144 container died 0edab4b746755a89fd50dba84ce83e2c1b7a081b5493524519b348bf2269ffbb (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:05.571 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.a.service' 2026-03-06T13:53:05.605 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:05.605 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-06T13:53:05.605 INFO:tasks.cephadm.mon.b:Stopping mon.b... 2026-03-06T13:53:05.605 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local systemd[1]: Stopping Ceph mon.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[62140]: 2026-03-06T12:53:05.690+0000 7fcadb7a3640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false --debug_ms 10 (PID: 1) UID: 0 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b[62140]: 2026-03-06T12:53:05.690+0000 7fcadb7a3640 -1 mon.b@1(peon) e2 *** Got Signal Terminated *** 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local podman[63204]: 2026-03-06 13:53:05.756111379 +0100 CET m=+0.077790635 container died e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local podman[63204]: 2026-03-06 13:53:05.872611833 +0100 CET m=+0.194291089 container remove e3b9538bda4f83e295bc708b0683b6d66dde1d10b5a903e54ee87c651332686f (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552) 2026-03-06T13:53:05.877 INFO:journalctl@ceph.mon.b.vm01.stdout:Mar 06 13:53:05 vm01.local bash[63204]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mon-b 2026-03-06T13:53:05.927 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mon.b.service' 2026-03-06T13:53:05.965 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:05.965 INFO:tasks.cephadm.mon.b:Stopped mon.b 2026-03-06T13:53:05.965 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-06T13:53:05.965 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a 2026-03-06T13:53:06.237 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:53:06 vm00.local systemd[1]: Stopping Ceph mgr.a for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:06.237 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:53:06 vm00.local podman[72608]: 2026-03-06 13:53:06.095640366 +0100 CET m=+0.052181968 container died eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:06.237 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:53:06 vm00.local podman[72608]: 2026-03-06 13:53:06.229565181 +0100 CET m=+0.186106783 container remove eda71fa52c98715734ef9275c51f4b3fc6095b4e662a7ae6a0aee2084ea1ec2b (image=harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-3, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:53:06.237 INFO:journalctl@ceph.mgr.a.vm00.stdout:Mar 06 13:53:06 vm00.local bash[72608]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-a 2026-03-06T13:53:06.282 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.a.service' 2026-03-06T13:53:06.313 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:06.313 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-06T13:53:06.313 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-06T13:53:06.313 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.b 2026-03-06T13:53:06.602 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:53:06 vm01.local systemd[1]: Stopping Ceph mgr.b for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:06.602 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:53:06 vm01.local podman[63318]: 2026-03-06 13:53:06.433343264 +0100 CET m=+0.043581370 container died f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:53:06.602 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:53:06 vm01.local podman[63318]: 2026-03-06 13:53:06.555227827 +0100 CET m=+0.165465933 container remove f384d59b827ef3dd7a24d729409d82771b18ba1d8eadf645a145ac7f6504e17a (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:53:06.602 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:53:06 vm01.local bash[63318]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-mgr-b 2026-03-06T13:53:06.602 INFO:journalctl@ceph.mgr.b.vm01.stdout:Mar 06 13:53:06 vm01.local systemd[1]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.b.service: Main process exited, code=exited, status=143/n/a 2026-03-06T13:53:06.609 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@mgr.b.service' 2026-03-06T13:53:06.643 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:06.643 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-06T13:53:06.643 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-06T13:53:06.643 DEBUG:teuthology.orchestra.run.vm00:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.0 2026-03-06T13:53:06.958 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:06 vm00.local systemd[1]: Stopping Ceph osd.0 for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:06.958 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:06 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0[62596]: 2026-03-06T12:53:06.729+0000 7fd543d16640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:53:06.958 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:06 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0[62596]: 2026-03-06T12:53:06.729+0000 7fd543d16640 -1 osd.0 13 *** Got signal Terminated *** 2026-03-06T13:53:06.958 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:06 vm00.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0[62596]: 2026-03-06T12:53:06.729+0000 7fd543d16640 -1 osd.0 13 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:53:12.013 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:11 vm00.local podman[72723]: 2026-03-06 13:53:11.754163156 +0100 CET m=+5.037004482 container died 8a6ce7135ff839dbaa9b053e075bb8e547fcfccdcf4573ff5a1fd92c176b8b20 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:53:12.013 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:11 vm00.local podman[72723]: 2026-03-06 13:53:11.872872252 +0100 CET m=+5.155713588 container remove 8a6ce7135ff839dbaa9b053e075bb8e547fcfccdcf4573ff5a1fd92c176b8b20 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:53:12.013 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:11 vm00.local bash[72723]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0 2026-03-06T13:53:12.013 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:11 vm00.local podman[72799]: 2026-03-06 13:53:11.990854755 +0100 CET m=+0.013911138 container create 7c820df5f8f30a6ff1d4e4eca9ba01ae9ef61baeadab14f97f363d33584a43cd (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0-deactivate, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:12.370 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:12 vm00.local podman[72799]: 2026-03-06 13:53:12.023042945 +0100 CET m=+0.046099328 container init 7c820df5f8f30a6ff1d4e4eca9ba01ae9ef61baeadab14f97f363d33584a43cd (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0-deactivate, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:12.370 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:12 vm00.local podman[72799]: 2026-03-06 13:53:12.027668697 +0100 CET m=+0.050725080 container start 7c820df5f8f30a6ff1d4e4eca9ba01ae9ef61baeadab14f97f363d33584a43cd (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0-deactivate, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:53:12.370 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:12 vm00.local podman[72799]: 2026-03-06 13:53:12.03001493 +0100 CET m=+0.053071322 container attach 7c820df5f8f30a6ff1d4e4eca9ba01ae9ef61baeadab14f97f363d33584a43cd (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-0-deactivate, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-06T13:53:12.370 INFO:journalctl@ceph.osd.0.vm00.stdout:Mar 06 13:53:12 vm00.local podman[72799]: 2026-03-06 13:53:11.985195308 +0100 CET m=+0.008251681 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:53:12.517 DEBUG:teuthology.orchestra.run.vm00:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.0.service' 2026-03-06T13:53:12.549 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:12.549 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-06T13:53:12.549 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-06T13:53:12.549 DEBUG:teuthology.orchestra.run.vm01:> sudo systemctl stop ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.1 2026-03-06T13:53:13.040 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:12 vm01.local systemd[1]: Stopping Ceph osd.1 for e1eed43a-195a-11f1-ac1a-e73a715574a2... 2026-03-06T13:53:13.040 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:12 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:53:12.644+0000 7fe6e8885640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-06T13:53:13.040 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:12 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:53:12.644+0000 7fe6e8885640 -1 osd.1 13 *** Got signal Terminated *** 2026-03-06T13:53:13.040 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:12 vm01.local ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1[58437]: 2026-03-06T12:53:12.644+0000 7fe6e8885640 -1 osd.1 13 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-06T13:53:17.926 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63433]: 2026-03-06 13:53:17.673980361 +0100 CET m=+5.043151654 container died 7fe23993daa2fd588aa66efa90e7a27782c9f0de72d9f5a8ffc1a08c96377f60 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2) 2026-03-06T13:53:17.926 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63433]: 2026-03-06 13:53:17.789348742 +0100 CET m=+5.158520025 container remove 7fe23993daa2fd588aa66efa90e7a27782c9f0de72d9f5a8ffc1a08c96377f60 (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8) 2026-03-06T13:53:17.926 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local bash[63433]: ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1 2026-03-06T13:53:17.927 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63511]: 2026-03-06 13:53:17.904085748 +0100 CET m=+0.013484021 container create 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git) 2026-03-06T13:53:18.184 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63511]: 2026-03-06 13:53:17.936612505 +0100 CET m=+0.046010778 container init 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:18.184 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63511]: 2026-03-06 13:53:17.94502891 +0100 CET m=+0.054427184 container start 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:53:18.184 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63511]: 2026-03-06 13:53:17.945854876 +0100 CET m=+0.055253149 container attach 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:53:18.184 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:17 vm01.local podman[63511]: 2026-03-06 13:53:17.898526232 +0100 CET m=+0.007924515 image pull 306e97de47e91c2b4b24d3dc09be3b3a12039b078f343d91220102acc6628a68 harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b 2026-03-06T13:53:18.436 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:18 vm01.local conmon[63522]: conmon 7476b38c1a5f808c2bd9 : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf.scope/memory.events 2026-03-06T13:53:18.436 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:18 vm01.local podman[63511]: 2026-03-06 13:53:18.3097966 +0100 CET m=+0.419194873 container died 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default) 2026-03-06T13:53:18.436 INFO:journalctl@ceph.osd.1.vm01.stdout:Mar 06 13:53:18 vm01.local podman[63511]: 2026-03-06 13:53:18.432586267 +0100 CET m=+0.541984531 container remove 7476b38c1a5f808c2bd92e045a978fa198f75ab6f27aa68366023743e367efbf (image=harbor.clyso.com/custom-ceph/ceph/ceph@sha256:26363c7a4eea9ef5a0148afc7b2a22b6f486596d87a30c2a9fdcda5db3eca62b, name=ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2-osd-1-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.8, CEPH_GIT_REPO=https://github.com/irq0/ceph.git, CEPH_REF=19.2.3-47-gc24117fd552, CEPH_SHA1=c24117fd5525679b799527bc1bd1f1dd0a2db5e2, FROM_IMAGE=rockylinux:9) 2026-03-06T13:53:18.446 DEBUG:teuthology.orchestra.run.vm01:> sudo pkill -f 'journalctl -f -n 0 -u ceph-e1eed43a-195a-11f1-ac1a-e73a715574a2@osd.1.service' 2026-03-06T13:53:18.491 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-06T13:53:18.491 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-06T13:53:18.491 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 --force --keep-logs 2026-03-06T13:53:18.745 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:53:20.912 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 --force --keep-logs 2026-03-06T13:53:21.169 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:53:23.254 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:53:23.279 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-06T13:53:23.302 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-06T13:53:23.302 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm00/crash 2026-03-06T13:53:23.302 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash -- . 2026-03-06T13:53:23.342 INFO:teuthology.orchestra.run.vm00.stderr:tar: /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash: Cannot open: No such file or directory 2026-03-06T13:53:23.342 INFO:teuthology.orchestra.run.vm00.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:53:23.343 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm01/crash 2026-03-06T13:53:23.344 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash -- . 2026-03-06T13:53:23.366 INFO:teuthology.orchestra.run.vm01.stderr:tar: /var/lib/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/crash: Cannot open: No such file or directory 2026-03-06T13:53:23.366 INFO:teuthology.orchestra.run.vm01.stderr:tar: Error is not recoverable: exiting now 2026-03-06T13:53:23.367 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-06T13:53:23.367 DEBUG:teuthology.orchestra.run.vm00:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-06T13:53:23.409 INFO:tasks.cephadm:Compressing logs... 2026-03-06T13:53:23.409 DEBUG:teuthology.orchestra.run.vm00:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:53:23.451 DEBUG:teuthology.orchestra.run.vm01:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:53:23.471 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:53:23.471 INFO:teuthology.orchestra.run.vm00.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:53:23.472 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.a.log 2026-03-06T13:53:23.472 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log 2026-03-06T13:53:23.472 INFO:teuthology.orchestra.run.vm01.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-06T13:53:23.472 INFO:teuthology.orchestra.run.vm01.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-06T13:53:23.473 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log 2026-03-06T13:53:23.474 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.a.log: 87.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:53:23.474 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.b.log 2026-03-06T13:53:23.474 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log 2026-03-06T13:53:23.474 INFO:teuthology.orchestra.run.vm01.stderr: 83.3% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-06T13:53:23.474 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log: 85.9% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log.gz 2026-03-06T13:53:23.475 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log 2026-03-06T13:53:23.480 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.b.log: gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log 2026-03-06T13:53:23.480 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.a.log 2026-03-06T13:53:23.480 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log: 90.4% 92.7% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log.gz 2026-03-06T13:53:23.481 INFO:teuthology.orchestra.run.vm01.stderr: -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log.gz 2026-03-06T13:53:23.481 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log: 90.1% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.audit.log.gz 2026-03-06T13:53:23.481 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log 2026-03-06T13:53:23.481 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log 2026-03-06T13:53:23.481 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log: 85.7% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.log.gz 2026-03-06T13:53:23.482 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.b.log 2026-03-06T13:53:23.482 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log: 78.4% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log.gz 2026-03-06T13:53:23.482 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.1.log 2026-03-06T13:53:23.484 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.b.log: 90.9% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.b.log.gz 2026-03-06T13:53:23.487 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log 2026-03-06T13:53:23.487 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log: 80.0% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph.cephadm.log.gz 2026-03-06T13:53:23.491 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.0.log 2026-03-06T13:53:23.493 INFO:teuthology.orchestra.run.vm01.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.1.log: 93.8% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.1.log.gz 2026-03-06T13:53:23.504 INFO:teuthology.orchestra.run.vm00.stderr:/var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log: /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.0.log: 92.8% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-volume.log.gz 2026-03-06T13:53:23.508 INFO:teuthology.orchestra.run.vm01.stderr: 93.4% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.b.log.gz 2026-03-06T13:53:23.509 INFO:teuthology.orchestra.run.vm01.stderr: 2026-03-06T13:53:23.509 INFO:teuthology.orchestra.run.vm01.stderr:real 0m0.046s 2026-03-06T13:53:23.509 INFO:teuthology.orchestra.run.vm01.stderr:user 0m0.053s 2026-03-06T13:53:23.509 INFO:teuthology.orchestra.run.vm01.stderr:sys 0m0.017s 2026-03-06T13:53:23.514 INFO:teuthology.orchestra.run.vm00.stderr: 88.9% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mgr.a.log.gz 2026-03-06T13:53:23.516 INFO:teuthology.orchestra.run.vm00.stderr: 93.8% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-osd.0.log.gz 2026-03-06T13:53:23.573 INFO:teuthology.orchestra.run.vm00.stderr: 91.9% -- replaced with /var/log/ceph/e1eed43a-195a-11f1-ac1a-e73a715574a2/ceph-mon.a.log.gz 2026-03-06T13:53:23.575 INFO:teuthology.orchestra.run.vm00.stderr: 2026-03-06T13:53:23.575 INFO:teuthology.orchestra.run.vm00.stderr:real 0m0.113s 2026-03-06T13:53:23.575 INFO:teuthology.orchestra.run.vm00.stderr:user 0m0.139s 2026-03-06T13:53:23.575 INFO:teuthology.orchestra.run.vm00.stderr:sys 0m0.016s 2026-03-06T13:53:23.575 INFO:tasks.cephadm:Archiving logs... 2026-03-06T13:53:23.575 DEBUG:teuthology.misc:Transferring archived files from vm00:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm00/log 2026-03-06T13:53:23.575 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:53:23.648 DEBUG:teuthology.misc:Transferring archived files from vm01:/var/log/ceph to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm01/log 2026-03-06T13:53:23.648 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-06T13:53:23.675 INFO:tasks.cephadm:Removing cluster... 2026-03-06T13:53:23.675 DEBUG:teuthology.orchestra.run.vm00:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 --force 2026-03-06T13:53:23.938 INFO:teuthology.orchestra.run.vm00.stdout:Deleting cluster with fsid: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:53:24.136 DEBUG:teuthology.orchestra.run.vm01:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid e1eed43a-195a-11f1-ac1a-e73a715574a2 --force 2026-03-06T13:53:24.390 INFO:teuthology.orchestra.run.vm01.stdout:Deleting cluster with fsid: e1eed43a-195a-11f1-ac1a-e73a715574a2 2026-03-06T13:53:24.581 INFO:tasks.cephadm:Removing cephadm ... 2026-03-06T13:53:24.581 DEBUG:teuthology.orchestra.run.vm00:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:53:24.596 DEBUG:teuthology.orchestra.run.vm01:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-06T13:53:24.610 INFO:tasks.cephadm:Teardown complete 2026-03-06T13:53:24.610 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-06T13:53:24.612 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-06T13:53:24.612 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:53:24.638 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-06T13:53:24.680 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:53:24.680 DEBUG:teuthology.orchestra.run.vm00:> 2026-03-06T13:53:24.680 DEBUG:teuthology.orchestra.run.vm00:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:53:24.680 DEBUG:teuthology.orchestra.run.vm00:> sudo yum -y remove $d || true 2026-03-06T13:53:24.680 DEBUG:teuthology.orchestra.run.vm00:> done 2026-03-06T13:53:24.685 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-06T13:53:24.685 DEBUG:teuthology.orchestra.run.vm01:> 2026-03-06T13:53:24.685 DEBUG:teuthology.orchestra.run.vm01:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-06T13:53:24.685 DEBUG:teuthology.orchestra.run.vm01:> sudo yum -y remove $d || true 2026-03-06T13:53:24.685 DEBUG:teuthology.orchestra.run.vm01:> done 2026-03-06T13:53:24.861 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:24.861 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 39 M 2026-03-06T13:53:24.862 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:24.864 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:24.864 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 39 M 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:53:24.876 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:24.877 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:24.877 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 39 M 2026-03-06T13:53:24.877 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:24.877 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:24.879 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:24.879 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:24.892 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:24.892 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:24.906 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:24.921 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:53:24.926 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:24.929 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.937 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.942 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.942 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:24.943 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-06T13:53:24.943 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-06T13:53:24.943 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-06T13:53:24.943 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:24.946 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.951 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:24.954 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:24.968 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:25.016 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:25.016 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:25.032 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:25.032 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 1/2 2026-03-06T13:53:25.062 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.063 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout: ceph-radosgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.075 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:25.248 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Remove 3 Packages 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 211 M 2026-03-06T13:53:25.249 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:25.251 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:25.251 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:25.257 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:25.257 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 210 M 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 211 M 2026-03-06T13:53:25.258 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:25.260 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:25.260 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:25.273 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:25.273 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:25.282 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:25.283 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:25.333 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:25.339 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:25.341 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:53:25.343 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:25.349 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:25.352 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:53:25.355 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.367 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.413 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.413 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:25.413 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:53:25.432 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.432 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:25.432 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.459 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: ceph-test-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.484 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:25.646 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Remove 8 Packages 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 28 M 2026-03-06T13:53:25.647 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:25.649 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:25.649 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:25.670 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:25.671 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:25.675 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: ceph x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 0 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 7.4 M 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 18 M 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-06T13:53:25.676 INFO:teuthology.orchestra.run.vm01.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout:Remove 8 Packages 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 28 M 2026-03-06T13:53:25.677 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:25.679 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:25.679 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:25.700 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:25.700 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:25.708 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:25.713 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:53:25.716 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:53:25.718 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:53:25.720 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:53:25.723 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:53:25.725 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:53:25.739 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.744 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:53:25.745 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.747 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-06T13:53:25.748 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-06T13:53:25.751 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.751 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-06T13:53:25.754 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-06T13:53:25.756 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-06T13:53:25.769 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.770 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:25.770 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:53:25.770 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:53:25.770 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:53:25.770 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.771 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-06T13:53:25.775 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.776 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.782 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 7/8 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-06T13:53:25.801 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.803 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.850 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.850 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:53:25.850 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:53:25.850 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:53:25.850 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:53:25.851 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:53:25.851 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:53:25.851 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-06T13:53:25.887 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-06T13:53:25.899 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:53:25.899 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.899 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:25.899 INFO:teuthology.orchestra.run.vm00.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:25.900 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: ceph-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mds-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mon-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: lua-5.4.4-4.el9.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: unzip-6.0-59.el9.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: zip-3.0-35.el9.x86_64 2026-03-06T13:53:25.940 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:25.941 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:26.097 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:53:26.103 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:53:26.104 INFO:teuthology.orchestra.run.vm00.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout:=================================================================================================== 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout:Remove 113 Packages 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 623 M 2026-03-06T13:53:26.105 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:26.131 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:26.132 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:26.132 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:26.138 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 23 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 431 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.4 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 803 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 88 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 66 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 563 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 59 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.4 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 85 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 626 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 60 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 1.5 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 51 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 138 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 425 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.6 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-06T13:53:26.139 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 702 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-06T13:53:26.140 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools x86_64 1:7.2-10.el9 @baseos 1.9 M 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout:=================================================================================================== 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout:Remove 113 Packages 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 623 M 2026-03-06T13:53:26.141 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:26.166 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:26.167 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:26.234 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:26.234 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:26.269 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:26.269 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:26.381 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:26.381 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:53:26.388 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:53:26.405 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.406 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.417 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:26.417 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:53:26.417 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.424 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 1/113 2026-03-06T13:53:26.439 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:53:26.439 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-06T13:53:26.443 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.444 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.451 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:53:26.455 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:53:26.455 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:53:26.456 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 2/113 2026-03-06T13:53:26.465 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:53:26.471 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:53:26.475 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:53:26.479 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 3/113 2026-03-06T13:53:26.479 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:53:26.483 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:53:26.487 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:53:26.490 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 4/113 2026-03-06T13:53:26.494 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/113 2026-03-06T13:53:26.494 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:53:26.504 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 6/113 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:53:26.506 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.510 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 7/113 2026-03-06T13:53:26.511 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.514 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 8/113 2026-03-06T13:53:26.519 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.521 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/113 2026-03-06T13:53:26.525 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/113 2026-03-06T13:53:26.535 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.535 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.535 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:53:26.535 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.542 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-06T13:53:26.545 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.550 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.550 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.553 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:53:26.557 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:53:26.558 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 11/113 2026-03-06T13:53:26.561 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:53:26.572 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:53:26.573 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.573 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.573 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-06T13:53:26.573 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.580 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.588 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 12/113 2026-03-06T13:53:26.591 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/113 2026-03-06T13:53:26.595 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/113 2026-03-06T13:53:26.599 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 15/113 2026-03-06T13:53:26.601 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:53:26.607 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:53:26.610 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:53:26.610 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/113 2026-03-06T13:53:26.619 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:53:26.630 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:53:26.630 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:53:26.638 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:53:26.639 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 17/113 2026-03-06T13:53:26.646 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 18/113 2026-03-06T13:53:26.648 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 19/113 2026-03-06T13:53:26.656 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 20/113 2026-03-06T13:53:26.667 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 21/113 2026-03-06T13:53:26.667 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:53:26.673 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 22/113 2026-03-06T13:53:26.728 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:53:26.742 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:53:26.748 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:53:26.752 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:53:26.761 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 23/113 2026-03-06T13:53:26.764 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.764 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:53:26.764 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.765 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.775 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 24/113 2026-03-06T13:53:26.781 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 25/113 2026-03-06T13:53:26.785 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 26/113 2026-03-06T13:53:26.789 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.794 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:53:26.796 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:53:26.797 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.797 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-06T13:53:26.797 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.798 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.808 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:53:26.814 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:53:26.816 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:53:26.818 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:53:26.822 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 27/113 2026-03-06T13:53:26.827 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 28/113 2026-03-06T13:53:26.829 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 29/113 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:53:26.838 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:26.839 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.841 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 30/113 2026-03-06T13:53:26.846 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 31/113 2026-03-06T13:53:26.848 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 32/113 2026-03-06T13:53:26.850 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.850 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 33/113 2026-03-06T13:53:26.854 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:53:26.856 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:53:26.858 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:53:26.861 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:53:26.864 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:53:26.867 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:53:26.867 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-06T13:53:26.871 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:26.873 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.885 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 34/113 2026-03-06T13:53:26.889 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 35/113 2026-03-06T13:53:26.891 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 36/113 2026-03-06T13:53:26.893 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 37/113 2026-03-06T13:53:26.896 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 38/113 2026-03-06T13:53:26.899 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 39/113 2026-03-06T13:53:26.902 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 40/113 2026-03-06T13:53:26.902 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:53:26.918 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:53:26.925 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:53:26.929 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:53:26.937 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:53:26.941 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:53:26.951 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:53:26.954 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 41/113 2026-03-06T13:53:26.956 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:53:26.961 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:53:26.961 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 42/113 2026-03-06T13:53:26.965 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:53:26.965 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 43/113 2026-03-06T13:53:26.973 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 44/113 2026-03-06T13:53:26.977 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 45/113 2026-03-06T13:53:26.987 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 46/113 2026-03-06T13:53:26.993 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 47/113 2026-03-06T13:53:26.998 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 48/113 2026-03-06T13:53:27.002 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 49/113 2026-03-06T13:53:27.010 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:53:27.020 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:53:27.023 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:53:27.027 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:53:27.029 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:53:27.033 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:53:27.035 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:53:27.047 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 50/113 2026-03-06T13:53:27.055 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.055 INFO:teuthology.orchestra.run.vm00.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:27.055 INFO:teuthology.orchestra.run.vm00.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:53:27.055 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:27.055 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.059 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 51/113 2026-03-06T13:53:27.062 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 52/113 2026-03-06T13:53:27.062 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.063 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:53:27.066 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:53:27.067 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 53/113 2026-03-06T13:53:27.068 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:53:27.069 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 54/113 2026-03-06T13:53:27.071 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:53:27.073 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 55/113 2026-03-06T13:53:27.076 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 56/113 2026-03-06T13:53:27.076 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:53:27.079 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:53:27.086 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:53:27.093 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:53:27.096 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.096 INFO:teuthology.orchestra.run.vm01.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-06T13:53:27.097 INFO:teuthology.orchestra.run.vm01.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-06T13:53:27.097 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:27.097 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.098 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:53:27.101 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:53:27.104 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:53:27.104 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-47.gc24117f 57/113 2026-03-06T13:53:27.106 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:53:27.106 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 58/113 2026-03-06T13:53:27.108 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:53:27.109 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 59/113 2026-03-06T13:53:27.110 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:53:27.112 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ply-3.11-14.el9.noarch 60/113 2026-03-06T13:53:27.113 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:53:27.115 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 61/113 2026-03-06T13:53:27.116 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:53:27.120 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 62/113 2026-03-06T13:53:27.122 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:53:27.124 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 63/113 2026-03-06T13:53:27.127 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:53:27.128 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:53:27.130 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 64/113 2026-03-06T13:53:27.131 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:53:27.136 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:53:27.139 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 65/113 2026-03-06T13:53:27.140 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:53:27.143 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:53:27.144 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 66/113 2026-03-06T13:53:27.146 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:53:27.147 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 67/113 2026-03-06T13:53:27.148 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:53:27.150 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 68/113 2026-03-06T13:53:27.153 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 69/113 2026-03-06T13:53:27.154 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:53:27.155 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 70/113 2026-03-06T13:53:27.157 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:53:27.158 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 71/113 2026-03-06T13:53:27.160 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:53:27.161 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 72/113 2026-03-06T13:53:27.162 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:53:27.163 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:53:27.164 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 73/113 2026-03-06T13:53:27.169 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:53:27.171 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/113 2026-03-06T13:53:27.172 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:53:27.176 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 75/113 2026-03-06T13:53:27.178 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 76/113 2026-03-06T13:53:27.182 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/113 2026-03-06T13:53:27.186 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 78/113 2026-03-06T13:53:27.189 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.189 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:53:27.189 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:27.191 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 79/113 2026-03-06T13:53:27.194 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 80/113 2026-03-06T13:53:27.196 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.198 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 81/113 2026-03-06T13:53:27.201 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 82/113 2026-03-06T13:53:27.206 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 83/113 2026-03-06T13:53:27.210 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:53:27.214 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 85/113 2026-03-06T13:53:27.217 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 86/113 2026-03-06T13:53:27.219 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 87/113 2026-03-06T13:53:27.223 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.223 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:53:27.225 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 88/113 2026-03-06T13:53:27.228 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 89/113 2026-03-06T13:53:27.235 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:53:27.240 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:53:27.243 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:53:27.245 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:53:27.245 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:53:27.246 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.246 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-06T13:53:27.246 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:27.253 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.279 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 90/113 2026-03-06T13:53:27.279 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:53:27.289 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 91/113 2026-03-06T13:53:27.294 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 92/113 2026-03-06T13:53:27.297 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 93/113 2026-03-06T13:53:27.300 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 94/113 2026-03-06T13:53:27.300 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /sys 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /proc 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /mnt 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /var/tmp 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /home 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /root 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout:skipping the directory /tmp 2026-03-06T13:53:32.313 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 95/113 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /sys 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /proc 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /mnt 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /var/tmp 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /home 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /root 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout:skipping the directory /tmp 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.320 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:53:32.331 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 96/113 2026-03-06T13:53:32.337 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.337 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.344 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.346 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.346 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.347 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:53:32.349 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:53:32.351 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:53:32.353 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:53:32.353 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:53:32.354 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 97/113 2026-03-06T13:53:32.357 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 98/113 2026-03-06T13:53:32.359 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 99/113 2026-03-06T13:53:32.362 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 100/113 2026-03-06T13:53:32.365 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 101/113 2026-03-06T13:53:32.365 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:53:32.365 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:53:32.376 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.376 INFO:teuthology.orchestra.run.vm00.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:53:32.376 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.377 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 102/113 2026-03-06T13:53:32.378 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.385 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.388 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:53:32.388 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.388 INFO:teuthology.orchestra.run.vm01.stdout:Removed "/etc/systemd/system/multi-user.target.wants/smartd.service". 2026-03-06T13:53:32.388 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.390 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.391 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:53:32.394 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:53:32.397 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:53:32.398 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: smartmontools-1:7.2-10.el9.x86_64 103/113 2026-03-06T13:53:32.400 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:53:32.401 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 104/113 2026-03-06T13:53:32.403 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 105/113 2026-03-06T13:53:32.406 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:53:32.406 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 106/113 2026-03-06T13:53:32.410 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 107/113 2026-03-06T13:53:32.413 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 108/113 2026-03-06T13:53:32.414 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:53:32.418 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 109/113 2026-03-06T13:53:32.419 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:53:32.422 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:53:32.422 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:53:32.425 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 110/113 2026-03-06T13:53:32.430 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 111/113 2026-03-06T13:53:32.433 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-msgpack-1.0.3-2.el9.x86_64 112/113 2026-03-06T13:53:32.433 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:53:32.526 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:53:32.527 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:53:32.528 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 113/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86 2/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x 3/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552 4/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-47.gc24117f 5/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 6/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.cl 7/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9. 8/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-47.gc2411 9/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:53:32.529 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9. 10/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.e 11/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso 12/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_ 13/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552. 14/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso. 15/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.n 16/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 17/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 18/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 19/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 20/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 21/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 22/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 23/113 2026-03-06T13:53:32.531 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso 24/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 25/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 26/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 27/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 28/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libradosstriper1-2:19.2.3-47.gc24117fd552.el9.cl 29/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 30/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 31/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 32/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 33/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 34/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 35/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 36/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 37/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 38/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 39/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 40/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 41/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 42/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 43/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-common-2:19.2.3-47.gc24117fd552.el9 44/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 45/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 46/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 47/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 48/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 49/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 50/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 51/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 52/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 53/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 54/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 55/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 56/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 57/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 58/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 59/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 60/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 61/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 62/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 63/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 64/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 65/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 66/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 67/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 68/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 69/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 70/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 71/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 72/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 73/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 74/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 75/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 76/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 77/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 78/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 79/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 80/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ply-3.11-14.el9.noarch 81/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 82/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 83/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 84/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 85/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 86/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 87/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 88/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 89/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 90/113 2026-03-06T13:53:32.532 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 91/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 92/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 93/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 94/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 95/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 96/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 97/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 98/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 99/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 100/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 101/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 102/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 103/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 104/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 105/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 106/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 107/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 108/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 109/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x8 110/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : smartmontools-1:7.2-10.el9.x86_64 111/113 2026-03-06T13:53:32.533 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 112/113 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:53:32.607 INFO:teuthology.orchestra.run.vm00.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:53:32.608 INFO:teuthology.orchestra.run.vm00.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:53:32.609 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 113/113 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-base-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-grafana-dashboards-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-immutable-object-cache-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-dashboard-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-diskprediction-local-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-k8sevents-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-modules-core-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-mgr-rook-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-osd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-prometheus-alerts-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-selinux-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ceph-volume-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libcephsqlite-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libradosstriper1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:53:32.610 INFO:teuthology.orchestra.run.vm01.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-common-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-06T13:53:32.611 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-ply-3.11-14.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-saml-1.16.0-1.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: rbd-mirror-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: smartmontools-1:7.2-10.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.612 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Remove 1 Package 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 775 k 2026-03-06T13:53:32.800 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:32.802 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:32.802 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout: cephadm noarch 2:19.2.3-47.gc24117fd552.el9.clyso @ceph-noarch 775 k 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:32.803 INFO:teuthology.orchestra.run.vm00.stdout:Remove 1 Package 2026-03-06T13:53:32.804 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.804 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 775 k 2026-03-06T13:53:32.804 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:32.805 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:32.805 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:32.806 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:32.807 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:32.819 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:32.819 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.822 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:32.822 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.927 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.932 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:32.964 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 1/1 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout: cephadm-2:19.2.3-47.gc24117fd552.el9.clyso.noarch 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:32.974 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.124 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:53:33.124 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.126 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.126 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.126 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.143 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-immutable-object-cache 2026-03-06T13:53:33.143 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.145 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.145 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.146 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.276 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr 2026-03-06T13:53:33.276 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.277 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.278 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.278 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.297 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr 2026-03-06T13:53:33.298 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.299 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.300 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.300 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.430 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:53:33.430 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.432 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.432 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.432 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.449 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-dashboard 2026-03-06T13:53:33.449 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.451 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.452 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.452 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.584 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:53:33.584 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.586 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.586 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.586 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.601 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-06T13:53:33.601 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.603 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.603 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.603 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.736 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:53:33.736 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.738 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.738 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.738 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.754 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-rook 2026-03-06T13:53:33.754 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.756 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.757 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.757 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:33.887 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:53:33.887 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:33.889 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:33.889 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:33.889 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:33.907 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-mgr-cephadm 2026-03-06T13:53:33.907 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:33.909 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:33.909 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:33.910 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:34.048 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:34.048 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 3.8 M 2026-03-06T13:53:34.049 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:34.051 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:34.051 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:34.063 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:34.063 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:34.069 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.6 M 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout: fuse x86_64 2.9.9-17.el9 @baseos 214 k 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.8 M 2026-03-06T13:53:34.070 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:34.072 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:34.072 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:34.085 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:34.085 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:34.090 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:34.093 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:53:34.105 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.112 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:34.115 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:53:34.128 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.171 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.171 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:53:34.183 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.183 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/2 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.212 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 2/2 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout: ceph-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 fuse-2.9.9-17.el9.x86_64 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.224 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:34.381 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: ceph-volume 2026-03-06T13:53:34.381 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:34.383 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:34.383 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:34.383 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:34.404 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: ceph-volume 2026-03-06T13:53:34.404 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:34.406 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:34.407 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:34.407 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:34.543 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Remove 2 Packages 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 610 k 2026-03-06T13:53:34.544 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:34.546 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:34.546 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:34.555 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:34.555 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:34.565 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 456 k 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 153 k 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Remove 2 Packages 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 610 k 2026-03-06T13:53:34.566 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:34.567 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:34.568 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:34.577 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:34.577 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:34.579 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:34.581 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:53:34.593 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.601 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:34.603 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:53:34.615 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.661 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.661 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:53:34.671 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.671 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x 1/2 2026-03-06T13:53:34.697 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.698 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:34.716 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x8 2/2 2026-03-06T13:53:34.716 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.716 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:34.716 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:34.716 INFO:teuthology.orchestra.run.vm01.stdout: librados-devel-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:34.717 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.717 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:34.868 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repo Size 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Remove 3 Packages 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 3.7 M 2026-03-06T13:53:34.869 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:34.871 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:34.871 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:34.886 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:34.886 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repo Size 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 3.0 M 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 514 k 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 187 k 2026-03-06T13:53:34.914 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout:Remove 3 Packages 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 3.7 M 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:34.915 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:34.916 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:34.916 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:34.916 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:53:34.918 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:53:34.918 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:53:34.930 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:34.930 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:34.960 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:34.962 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 1/3 2026-03-06T13:53:34.964 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:53:34.964 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:53:34.973 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:53:34.973 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:34.973 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.007 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:35.021 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 3/3 2026-03-06T13:53:35.021 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/3 2026-03-06T13:53:35.021 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.c 2/3 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x8 3/3 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: libcephfs2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: python3-ceph-argparse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: python3-cephfs-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.056 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:35.161 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: libcephfs-devel 2026-03-06T13:53:35.161 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:35.163 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:35.163 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:35.163 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:35.209 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: libcephfs-devel 2026-03-06T13:53:35.209 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:35.211 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:35.212 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:35.212 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:35.322 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:35.323 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:35.323 INFO:teuthology.orchestra.run.vm00.stdout: Package Arch Version Repository Size 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Removing: 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Removing dependent packages: 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Removing unused dependencies: 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Transaction Summary 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:================================================================================ 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Remove 20 Packages 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Freed space: 79 M 2026-03-06T13:53:35.324 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction check 2026-03-06T13:53:35.328 INFO:teuthology.orchestra.run.vm00.stdout:Transaction check succeeded. 2026-03-06T13:53:35.328 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction test 2026-03-06T13:53:35.349 INFO:teuthology.orchestra.run.vm00.stdout:Transaction test succeeded. 2026-03-06T13:53:35.349 INFO:teuthology.orchestra.run.vm00.stdout:Running transaction 2026-03-06T13:53:35.370 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: Package Arch Version Repository Size 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Removing: 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: librados2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Removing dependent packages: 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 1.1 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 265 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 227 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 494 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Removing unused dependencies: 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: librbd1 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 13 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: librgw2 x86_64 2:19.2.3-47.gc24117fd552.el9.clyso @ceph 19 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Transaction Summary 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:================================================================================ 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Remove 20 Packages 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Freed space: 79 M 2026-03-06T13:53:35.372 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction check 2026-03-06T13:53:35.376 INFO:teuthology.orchestra.run.vm01.stdout:Transaction check succeeded. 2026-03-06T13:53:35.376 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction test 2026-03-06T13:53:35.387 INFO:teuthology.orchestra.run.vm00.stdout: Preparing : 1/1 2026-03-06T13:53:35.390 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:53:35.392 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:53:35.395 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:53:35.395 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:53:35.396 INFO:teuthology.orchestra.run.vm01.stdout:Transaction test succeeded. 2026-03-06T13:53:35.396 INFO:teuthology.orchestra.run.vm01.stdout:Running transaction 2026-03-06T13:53:35.406 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:53:35.408 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:53:35.410 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:53:35.411 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:53:35.413 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:53:35.415 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:53:35.416 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.428 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.428 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:53:35.428 INFO:teuthology.orchestra.run.vm00.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:53:35.428 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.435 INFO:teuthology.orchestra.run.vm01.stdout: Preparing : 1/1 2026-03-06T13:53:35.437 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 1/20 2026-03-06T13:53:35.439 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2/20 2026-03-06T13:53:35.440 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:53:35.442 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:53:35.442 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 3/20 2026-03-06T13:53:35.442 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:53:35.445 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:53:35.448 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:53:35.450 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:53:35.453 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:53:35.454 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 4/20 2026-03-06T13:53:35.455 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:53:35.456 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-06T13:53:35.456 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:53:35.457 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 6/20 2026-03-06T13:53:35.458 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:53:35.459 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 7/20 2026-03-06T13:53:35.460 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-06T13:53:35.463 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-06T13:53:35.463 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.471 INFO:teuthology.orchestra.run.vm00.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:53:35.475 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.475 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:53:35.475 INFO:teuthology.orchestra.run.vm01.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-06T13:53:35.475 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.487 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 11/20 2026-03-06T13:53:35.489 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:53:35.492 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-06T13:53:35.495 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-06T13:53:35.497 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-06T13:53:35.500 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-06T13:53:35.502 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-06T13:53:35.504 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-06T13:53:35.506 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-06T13:53:35.519 INFO:teuthology.orchestra.run.vm01.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:53:35.526 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:53:35.527 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout:Removed: 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout: 2026-03-06T13:53:35.577 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_6 7/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 8/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 10/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-06T13:53:35.578 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x 13/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86 14/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86 15/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 17/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 18/20 2026-03-06T13:53:35.579 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout:Removed: 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: librados2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: librbd1-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: librgw2-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: python3-rados-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: python3-rbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: python3-rgw-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: rbd-fuse-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: rbd-nbd-2:19.2.3-47.gc24117fd552.el9.clyso.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: re2-1:20211101-20.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout: 2026-03-06T13:53:35.628 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:35.762 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: librbd1 2026-03-06T13:53:35.762 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:35.764 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:35.765 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:35.765 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:35.823 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: librbd1 2026-03-06T13:53:35.823 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:35.826 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:35.826 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:35.826 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:35.941 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rados 2026-03-06T13:53:35.941 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:35.943 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:35.943 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:35.943 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:35.998 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rados 2026-03-06T13:53:35.998 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.000 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.001 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.001 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.092 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rgw 2026-03-06T13:53:36.092 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.094 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.095 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.095 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.149 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rgw 2026-03-06T13:53:36.149 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.151 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.152 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.152 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.243 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-cephfs 2026-03-06T13:53:36.243 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.245 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.245 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.245 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.299 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-cephfs 2026-03-06T13:53:36.300 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.301 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.302 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.302 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.392 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: python3-rbd 2026-03-06T13:53:36.392 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.394 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.394 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.394 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.450 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: python3-rbd 2026-03-06T13:53:36.450 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.452 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.453 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.453 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.545 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-fuse 2026-03-06T13:53:36.545 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.547 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.548 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.548 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.600 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-fuse 2026-03-06T13:53:36.601 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.603 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.603 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.603 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.695 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-mirror 2026-03-06T13:53:36.695 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.697 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.698 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.698 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.751 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-mirror 2026-03-06T13:53:36.752 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.753 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.754 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.754 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.845 INFO:teuthology.orchestra.run.vm00.stdout:No match for argument: rbd-nbd 2026-03-06T13:53:36.846 INFO:teuthology.orchestra.run.vm00.stderr:No packages marked for removal. 2026-03-06T13:53:36.847 INFO:teuthology.orchestra.run.vm00.stdout:Dependencies resolved. 2026-03-06T13:53:36.848 INFO:teuthology.orchestra.run.vm00.stdout:Nothing to do. 2026-03-06T13:53:36.848 INFO:teuthology.orchestra.run.vm00.stdout:Complete! 2026-03-06T13:53:36.867 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean all 2026-03-06T13:53:36.902 INFO:teuthology.orchestra.run.vm01.stdout:No match for argument: rbd-nbd 2026-03-06T13:53:36.902 INFO:teuthology.orchestra.run.vm01.stderr:No packages marked for removal. 2026-03-06T13:53:36.904 INFO:teuthology.orchestra.run.vm01.stdout:Dependencies resolved. 2026-03-06T13:53:36.904 INFO:teuthology.orchestra.run.vm01.stdout:Nothing to do. 2026-03-06T13:53:36.904 INFO:teuthology.orchestra.run.vm01.stdout:Complete! 2026-03-06T13:53:36.923 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean all 2026-03-06T13:53:36.985 INFO:teuthology.orchestra.run.vm00.stdout:56 files removed 2026-03-06T13:53:37.004 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:53:37.026 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:53:37.040 INFO:teuthology.orchestra.run.vm01.stdout:56 files removed 2026-03-06T13:53:37.057 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph-source.repo 2026-03-06T13:53:37.078 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph-noarch.repo 2026-03-06T13:53:37.088 DEBUG:teuthology.orchestra.run.vm00:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:53:37.143 DEBUG:teuthology.orchestra.run.vm01:> sudo rm /etc/yum.repos.d/ceph.repo 2026-03-06T13:53:37.150 DEBUG:teuthology.orchestra.run.vm00:> sudo yum clean expire-cache 2026-03-06T13:53:37.207 DEBUG:teuthology.orchestra.run.vm01:> sudo yum clean expire-cache 2026-03-06T13:53:37.290 INFO:teuthology.orchestra.run.vm00.stdout:Cache was expired 2026-03-06T13:53:37.290 INFO:teuthology.orchestra.run.vm00.stdout:0 files removed 2026-03-06T13:53:37.306 DEBUG:teuthology.parallel:result is None 2026-03-06T13:53:37.350 INFO:teuthology.orchestra.run.vm01.stdout:Cache was expired 2026-03-06T13:53:37.350 INFO:teuthology.orchestra.run.vm01.stdout:0 files removed 2026-03-06T13:53:37.365 DEBUG:teuthology.parallel:result is None 2026-03-06T13:53:37.365 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm00.local 2026-03-06T13:53:37.365 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm01.local 2026-03-06T13:53:37.365 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:53:37.365 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-06T13:53:37.387 DEBUG:teuthology.orchestra.run.vm00:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:53:37.388 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-06T13:53:37.448 INFO:teuthology.orchestra.run.vm00.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:53:37.449 INFO:teuthology.orchestra.run.vm01.stderr:mv: cannot stat '/etc/yum/pluginconf.d/priorities.conf.orig': No such file or directory 2026-03-06T13:53:37.449 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:53:37.449 DEBUG:teuthology.parallel:result is None 2026-03-06T13:53:37.449 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:53:37.450 DEBUG:teuthology.parallel:result is None 2026-03-06T13:53:37.450 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-06T13:53:37.452 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-06T13:53:37.452 DEBUG:teuthology.orchestra.run.vm00:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:53:37.491 DEBUG:teuthology.orchestra.run.vm01:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-06T13:53:37.502 INFO:teuthology.orchestra.run.vm00.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:53:37.503 INFO:teuthology.orchestra.run.vm01.stderr:bash: line 1: ntpq: command not found 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:=============================================================================== 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:^* ntp01.pingless.com 2 6 377 53 +635us[ +567us] +/- 14ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:^+ 172-236-195-26.ip.linode> 3 6 377 54 +1148us[+1079us] +/- 20ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:^+ sv1.ggsrv.de 2 6 377 54 -108us[ -176us] +/- 18ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm01.stdout:^+ ntp.kernfusion.at 2 6 377 53 -3010us[-3010us] +/- 28ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:=============================================================================== 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:^* ntp01.pingless.com 2 6 377 54 +839us[ +864us] +/- 14ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:^+ 172-236-195-26.ip.linode> 3 6 377 118 +1178us[+1306us] +/- 20ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:^+ sv1.ggsrv.de 2 6 377 53 +44us[ +44us] +/- 18ms 2026-03-06T13:53:37.647 INFO:teuthology.orchestra.run.vm00.stdout:^+ ntp.kernfusion.at 2 6 377 53 -3064us[-3064us] +/- 28ms 2026-03-06T13:53:37.647 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-06T13:53:37.649 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-06T13:53:37.650 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-06T13:53:37.652 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-06T13:53:37.654 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-06T13:53:37.655 INFO:teuthology.task.internal:Duration was 446.588273 seconds 2026-03-06T13:53:37.656 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-06T13:53:37.658 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-06T13:53:37.658 DEBUG:teuthology.orchestra.run.vm00:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:53:37.689 DEBUG:teuthology.orchestra.run.vm01:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-06T13:53:37.723 INFO:teuthology.orchestra.run.vm01.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:53:37.726 INFO:teuthology.orchestra.run.vm00.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-06T13:53:38.236 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-06T13:53:38.236 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm00.local 2026-03-06T13:53:38.236 DEBUG:teuthology.orchestra.run.vm00:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:53:38.298 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm01.local 2026-03-06T13:53:38.298 DEBUG:teuthology.orchestra.run.vm01:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-06T13:53:38.320 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-06T13:53:38.320 DEBUG:teuthology.orchestra.run.vm00:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:53:38.340 DEBUG:teuthology.orchestra.run.vm01:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:53:38.751 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-06T13:53:38.751 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:53:38.753 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-06T13:53:38.774 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:53:38.775 INFO:teuthology.orchestra.run.vm00.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:53:38.775 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:53:38.775 INFO:teuthology.orchestra.run.vm00.stderr: -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:53:38.775 INFO:teuthology.orchestra.run.vm00.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:53:38.776 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-06T13:53:38.776 INFO:teuthology.orchestra.run.vm01.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-06T13:53:38.777 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-06T13:53:38.777 INFO:teuthology.orchestra.run.vm01.stderr: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-06T13:53:38.777 INFO:teuthology.orchestra.run.vm01.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0%/home/ubuntu/cephtest/archive/syslog/journalctl.log: -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-06T13:53:38.905 INFO:teuthology.orchestra.run.vm01.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:53:38.912 INFO:teuthology.orchestra.run.vm00.stderr: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-06T13:53:38.914 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-06T13:53:38.917 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-06T13:53:38.917 DEBUG:teuthology.orchestra.run.vm00:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:53:38.977 DEBUG:teuthology.orchestra.run.vm01:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-06T13:53:39.002 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-06T13:53:39.004 DEBUG:teuthology.orchestra.run.vm00:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:53:39.019 DEBUG:teuthology.orchestra.run.vm01:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:53:39.040 INFO:teuthology.orchestra.run.vm00.stdout:kernel.core_pattern = core 2026-03-06T13:53:39.070 INFO:teuthology.orchestra.run.vm01.stdout:kernel.core_pattern = core 2026-03-06T13:53:39.083 DEBUG:teuthology.orchestra.run.vm00:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:53:39.105 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:53:39.105 DEBUG:teuthology.orchestra.run.vm01:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-06T13:53:39.139 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-06T13:53:39.139 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-06T13:53:39.141 INFO:teuthology.task.internal:Transferring archived files... 2026-03-06T13:53:39.142 DEBUG:teuthology.misc:Transferring archived files from vm00:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm00 2026-03-06T13:53:39.142 DEBUG:teuthology.orchestra.run.vm00:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:53:39.170 DEBUG:teuthology.misc:Transferring archived files from vm01:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-06_13:20:18-orch:cephadm:workunits-cobaltcore-storage-v19.2.3-fasttrack-3-none-default-vps/278/remote/vm01 2026-03-06T13:53:39.170 DEBUG:teuthology.orchestra.run.vm01:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-06T13:53:39.207 INFO:teuthology.task.internal:Removing archive directory... 2026-03-06T13:53:39.207 DEBUG:teuthology.orchestra.run.vm00:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:53:39.211 DEBUG:teuthology.orchestra.run.vm01:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-06T13:53:39.263 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-06T13:53:39.266 INFO:teuthology.task.internal:Not uploading archives. 2026-03-06T13:53:39.266 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-06T13:53:39.268 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-06T13:53:39.268 DEBUG:teuthology.orchestra.run.vm00:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:53:39.269 DEBUG:teuthology.orchestra.run.vm01:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-06T13:53:39.282 INFO:teuthology.orchestra.run.vm00.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:53 /home/ubuntu/cephtest 2026-03-06T13:53:39.319 INFO:teuthology.orchestra.run.vm01.stdout: 8532147 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 6 13:53 /home/ubuntu/cephtest 2026-03-06T13:53:39.320 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-06T13:53:39.325 INFO:teuthology.run:Summary data: description: orch:cephadm:workunits/{0-distro/centos_9.stream_runc agent/on mon_election/connectivity task/test_extra_daemon_features} duration: 446.5882730484009 flavor: default owner: irq0 success: true 2026-03-06T13:53:39.325 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-06T13:53:39.343 INFO:teuthology.run:pass