2026-03-10T12:52:00.510 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T12:52:00.514 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T12:52:00.532 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037 branch: squid description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_host_drain} email: null first_in_suite: false flavor: default job_id: '1037' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - mons down - mon down - out of quorum - CEPHADM_STRAY_HOST - CEPHADM_STRAY_DAEMON - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - osd.1 - - host.b - mon.b - mgr.b - osd.2 - osd.3 - - host.c - mon.c - osd.4 - osd.5 seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEnKVCRaraPcOCGvxCTlWK+F5ihgxn2XsGiHmxKmaJ+2I30ag/jYBpMl5cumMEDFwA8fa9txmg3CkzMskuFHpjY= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBiKfFiF+aOZ7GGR404JsnjJoeozHJjcleacy5Je4dofNjk9iWm89fVHKSyGDgqLZRWsMaFztqHL9BT+3dKcT6w= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMkzYa74oLMCZP1npxv2u5uKVgNK8pTMEaDRuK5plIQlX+A6ZjvaN+/pcN1boQ0enxO9il0QEuBYItKFOaElWQ= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: null - cephadm.shell: host.a: - "set -ex\nHOSTNAMES=$(ceph orch host ls --format json | jq -r '.[] | .hostname')\n\ for host in $HOSTNAMES; do\n # find the hostname for \"host.c\" which will\ \ have no mgr\n HAS_MGRS=$(ceph orch ps --hostname ${host} --format json |\ \ jq 'any(.daemon_type == \"mgr\")')\n if [ \"$HAS_MGRS\" == \"false\" ]; then\n\ \ HOST_C=\"${host}\"\n fi\ndone\n# One last thing to worry about before\ \ draining the host\n# is that the teuthology test tends to put the explicit\n\ # hostnames in the placement for the mon service.\n# We want to make sure we\ \ can drain without providing\n# --force and there is a check for the host being\ \ removed\n# being listed explicitly in the placements. Therefore,\n# we should\ \ remove it from the mon placement.\nceph orch ls mon --export > mon.yaml\n\ sed /\"$HOST_C\"/d mon.yaml > mon_adjusted.yaml\nceph orch apply -i mon_adjusted.yaml\n\ # now drain that host\nceph orch host drain $HOST_C --zap-osd-devices\n# wait\ \ for drain to complete\nHOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C)\n\ while [ \"$HOST_C_DAEMONS\" != \"No daemons reported\" ]; do\n sleep 15\n \ \ HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C)\ndone\n# we want to check\ \ the ability to remove the host from\n# the CRUSH map, so we should first verify\ \ the host is in\n# the CRUSH map.\nceph osd getcrushmap -o compiled-crushmap\n\ crushtool -d compiled-crushmap -o crushmap.txt\nCRUSH_MAP=$(cat crushmap.txt)\n\ if ! grep -q \"$HOST_C\" <<< \"$CRUSH_MAP\"; then\n printf \"Expected to see\ \ $HOST_C in CRUSH map. Saw:\\n\\n$CRUSH_MAP\"\n exit 1\nfi\n# If the drain\ \ was successful, we should be able to remove the\n# host without force with\ \ no issues. If there are still daemons\n# we will get a response telling us\ \ to drain the host and a\n# non-zero return code\nceph orch host rm $HOST_C\ \ --rm-crush-entry\n# verify we've successfully removed the host from the CRUSH\ \ map\nsleep 30\nceph osd getcrushmap -o compiled-crushmap\ncrushtool -d compiled-crushmap\ \ -o crushmap.txt\nCRUSH_MAP=$(cat crushmap.txt)\nif grep -q \"$HOST_C\" <<<\ \ \"$CRUSH_MAP\"; then\n printf \"Saw $HOST_C in CRUSH map after it should\ \ have been removed.\\n\\n$CRUSH_MAP\"\n exit 1\nfi\n" teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T12:52:00.532 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T12:52:00.533 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T12:52:00.533 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T12:52:00.533 INFO:teuthology.task.internal:Checking packages... 2026-03-10T12:52:00.533 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T12:52:00.533 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T12:52:00.533 INFO:teuthology.packaging:ref: None 2026-03-10T12:52:00.533 INFO:teuthology.packaging:tag: None 2026-03-10T12:52:00.533 INFO:teuthology.packaging:branch: squid 2026-03-10T12:52:00.533 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:52:00.533 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T12:52:01.249 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T12:52:01.250 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T12:52:01.250 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T12:52:01.250 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T12:52:01.251 INFO:teuthology.task.internal:Saving configuration 2026-03-10T12:52:01.256 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T12:52:01.256 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T12:52:01.262 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm04.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 12:50:09.561112', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:04', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEnKVCRaraPcOCGvxCTlWK+F5ihgxn2XsGiHmxKmaJ+2I30ag/jYBpMl5cumMEDFwA8fa9txmg3CkzMskuFHpjY='} 2026-03-10T12:52:01.266 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 12:50:09.560898', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBiKfFiF+aOZ7GGR404JsnjJoeozHJjcleacy5Je4dofNjk9iWm89fVHKSyGDgqLZRWsMaFztqHL9BT+3dKcT6w='} 2026-03-10T12:52:01.271 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 12:50:09.560517', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMkzYa74oLMCZP1npxv2u5uKVgNK8pTMEaDRuK5plIQlX+A6ZjvaN+/pcN1boQ0enxO9il0QEuBYItKFOaElWQ='} 2026-03-10T12:52:01.271 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T12:52:01.272 INFO:teuthology.task.internal:roles: ubuntu@vm04.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0', 'osd.1'] 2026-03-10T12:52:01.272 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['host.b', 'mon.b', 'mgr.b', 'osd.2', 'osd.3'] 2026-03-10T12:52:01.272 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['host.c', 'mon.c', 'osd.4', 'osd.5'] 2026-03-10T12:52:01.272 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T12:52:01.278 DEBUG:teuthology.task.console_log:vm04 does not support IPMI; excluding 2026-03-10T12:52:01.282 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-03-10T12:52:01.287 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-10T12:52:01.287 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f15430cff40>, signals=[15]) 2026-03-10T12:52:01.287 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T12:52:01.288 INFO:teuthology.task.internal:Opening connections... 2026-03-10T12:52:01.288 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-10T12:52:01.288 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:52:01.345 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-10T12:52:01.346 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:52:01.403 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-10T12:52:01.403 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:52:01.463 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T12:52:01.464 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-10T12:52:01.481 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-10T12:52:01.481 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-10T12:52:01.535 INFO:teuthology.orchestra.run.vm04.stdout:NAME="CentOS Stream" 2026-03-10T12:52:01.535 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="9" 2026-03-10T12:52:01.535 INFO:teuthology.orchestra.run.vm04.stdout:ID="centos" 2026-03-10T12:52:01.535 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE="rhel fedora" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="9" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:PLATFORM_ID="platform:el9" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:ANSI_COLOR="0;31" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:LOGO="fedora-logo-icon" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://centos.org/" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T12:52:01.536 INFO:teuthology.orchestra.run.vm04.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T12:52:01.536 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-10T12:52:01.541 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-10T12:52:01.557 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-10T12:52:01.558 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T12:52:01.614 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T12:52:01.615 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-10T12:52:01.619 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-10T12:52:01.634 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-10T12:52:01.634 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:NAME="CentOS Stream" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:ID="centos" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE="rhel fedora" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:PLATFORM_ID="platform:el9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:ANSI_COLOR="0;31" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:LOGO="fedora-logo-icon" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://centos.org/" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T12:52:01.690 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T12:52:01.691 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-10T12:52:01.713 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T12:52:01.737 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T12:52:01.785 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T12:52:01.785 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-10T12:52:01.787 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-10T12:52:01.789 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-10T12:52:01.803 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T12:52:01.815 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T12:52:01.815 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-10T12:52:01.842 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-10T12:52:01.844 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-10T12:52:01.856 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T12:52:01.860 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T12:52:01.861 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T12:52:01.861 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T12:52:01.880 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-10T12:52:01.912 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.104 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-10T12:52:02.119 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.304 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-10T12:52:02.317 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.489 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T12:52:02.490 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T12:52:02.490 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T12:52:02.492 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T12:52:02.494 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T12:52:02.508 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T12:52:02.509 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T12:52:02.510 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T12:52:02.511 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T12:52:02.549 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T12:52:02.551 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T12:52:02.570 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T12:52:02.571 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T12:52:02.571 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T12:52:02.619 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.619 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T12:52:02.633 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.633 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T12:52:02.647 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:52:02.647 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T12:52:02.661 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T12:52:02.675 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T12:52:02.682 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.691 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.697 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.706 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.711 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.721 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T12:52:02.722 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T12:52:02.742 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T12:52:02.742 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T12:52:02.744 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T12:52:02.749 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T12:52:02.786 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T12:52:02.819 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T12:52:02.819 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T12:52:02.821 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T12:52:02.823 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T12:52:02.841 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:52:02.898 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:52:02.955 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:52:02.955 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T12:52:03.012 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:52:03.039 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:52:03.090 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:52:03.090 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T12:52:03.147 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T12:52:03.169 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T12:52:03.225 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:52:03.225 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T12:52:03.282 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-10T12:52:03.284 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-10T12:52:03.286 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-10T12:52:03.310 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T12:52:03.312 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T12:52:03.349 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T12:52:03.653 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T12:52:03.655 INFO:teuthology.task.internal:Starting timer... 2026-03-10T12:52:03.655 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T12:52:03.658 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T12:52:03.660 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-10T12:52:03.660 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-10T12:52:03.660 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-10T12:52:03.660 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-10T12:52:03.660 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T12:52:03.660 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T12:52:03.660 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T12:52:03.660 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T12:52:03.662 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T12:52:03.662 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T12:52:03.663 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T12:52:04.168 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T12:52:04.174 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T12:52:04.174 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryya6ljh8k --limit vm04.local,vm06.local,vm09.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T12:54:10.228 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm09.local')] 2026-03-10T12:54:10.229 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-10T12:54:10.229 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:54:10.295 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-10T12:54:10.382 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-10T12:54:10.382 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-10T12:54:10.382 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:54:10.448 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-10T12:54:10.527 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-10T12:54:10.527 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-10T12:54:10.527 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T12:54:10.590 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-10T12:54:10.669 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-10T12:54:10.670 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T12:54:10.672 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T12:54:10.672 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T12:54:10.672 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:54:10.676 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T12:54:10.676 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:54:10.678 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T12:54:10.678 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T12:54:10.714 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T12:54:10.723 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T12:54:10.731 INFO:teuthology.orchestra.run.vm04.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T12:54:10.743 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T12:54:10.747 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T12:54:10.754 INFO:teuthology.orchestra.run.vm04.stderr:sudo: ntpd: command not found 2026-03-10T12:54:10.764 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T12:54:10.765 INFO:teuthology.orchestra.run.vm04.stdout:506 Cannot talk to daemon 2026-03-10T12:54:10.778 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-10T12:54:10.779 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T12:54:10.791 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-10T12:54:10.792 INFO:teuthology.orchestra.run.vm09.stderr:sudo: ntpd: command not found 2026-03-10T12:54:10.793 INFO:teuthology.orchestra.run.vm04.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T12:54:10.805 INFO:teuthology.orchestra.run.vm09.stdout:506 Cannot talk to daemon 2026-03-10T12:54:10.810 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T12:54:10.819 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T12:54:10.823 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T12:54:10.833 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T12:54:10.837 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:^? ntp3.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:^? 141.144.246.224 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:^? web80.weingaertner-it.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.842 INFO:teuthology.orchestra.run.vm04.stdout:^? srv01-nc.securepod.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.876 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:^? srv01-nc.securepod.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:^? ntp3.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:^? 141.144.246.224 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.879 INFO:teuthology.orchestra.run.vm06.stdout:^? web80.weingaertner-it.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.881 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:^? web80.weingaertner-it.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:^? srv01-nc.securepod.org 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:^? ntp3.uni-ulm.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.884 INFO:teuthology.orchestra.run.vm09.stdout:^? 141.144.246.224 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-03-10T12:54:10.884 INFO:teuthology.run_tasks:Running task pexec... 2026-03-10T12:54:10.887 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-10T12:54:10.887 DEBUG:teuthology.orchestra.run.vm04:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T12:54:10.887 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T12:54:10.887 DEBUG:teuthology.orchestra.run.vm09:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T12:54:10.888 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.888 DEBUG:teuthology.task.pexec:ubuntu@vm04.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:10.888 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm04.local 2026-03-10T12:54:10.888 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.888 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:10.890 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.890 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:10.890 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-10T12:54:10.890 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.890 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:10.927 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.927 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:10.927 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm09.local 2026-03-10T12:54:10.927 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T12:54:10.927 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-10T12:54:11.082 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: nvme-cli 2026-03-10T12:54:11.082 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T12:54:11.085 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T12:54:11.086 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T12:54:11.086 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T12:54:11.146 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-10T12:54:11.146 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T12:54:11.147 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: nvme-cli 2026-03-10T12:54:11.147 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T12:54:11.149 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T12:54:11.150 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T12:54:11.150 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T12:54:11.151 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T12:54:11.151 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T12:54:11.152 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T12:54:11.497 INFO:teuthology.orchestra.run.vm04.stdout:Last metadata expiration check: 0:01:20 ago on Tue 10 Mar 2026 12:52:51 PM UTC. 2026-03-10T12:54:11.551 INFO:teuthology.orchestra.run.vm09.stdout:Last metadata expiration check: 0:01:34 ago on Tue 10 Mar 2026 12:52:37 PM UTC. 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: Package Architecture Version Repository Size 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T12:54:11.596 INFO:teuthology.orchestra.run.vm04.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:Install 6 Packages 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 2.3 M 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:Installed size: 11 M 2026-03-10T12:54:11.597 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-10T12:54:11.644 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:16 ago on Tue 10 Mar 2026 12:52:55 PM UTC. 2026-03-10T12:54:11.645 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: Package Architecture Version Repository Size 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Install 6 Packages 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 2.3 M 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Installed size: 11 M 2026-03-10T12:54:11.646 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T12:54:11.768 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: Package Architecture Version Repository Size 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Install 6 Packages 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 2.3 M 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 11 M 2026-03-10T12:54:11.772 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T12:54:12.084 INFO:teuthology.orchestra.run.vm06.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 224 kB/s | 44 kB 00:00 2026-03-10T12:54:12.108 INFO:teuthology.orchestra.run.vm06.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 328 kB/s | 72 kB 00:00 2026-03-10T12:54:12.183 INFO:teuthology.orchestra.run.vm06.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 858 kB/s | 84 kB 00:00 2026-03-10T12:54:12.219 INFO:teuthology.orchestra.run.vm06.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.3 MB/s | 150 kB 00:00 2026-03-10T12:54:12.303 INFO:teuthology.orchestra.run.vm04.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 172 kB/s | 44 kB 00:00 2026-03-10T12:54:12.306 INFO:teuthology.orchestra.run.vm04.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 279 kB/s | 72 kB 00:00 2026-03-10T12:54:12.307 INFO:teuthology.orchestra.run.vm09.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 345 kB/s | 72 kB 00:00 2026-03-10T12:54:12.322 INFO:teuthology.orchestra.run.vm09.stdout:(2/6): nvmetcli-0.8-3.el9.noarch.rpm 197 kB/s | 44 kB 00:00 2026-03-10T12:54:12.382 INFO:teuthology.orchestra.run.vm09.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.1 MB/s | 84 kB 00:00 2026-03-10T12:54:12.383 INFO:teuthology.orchestra.run.vm06.stdout:(5/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 4.1 MB/s | 837 kB 00:00 2026-03-10T12:54:12.432 INFO:teuthology.orchestra.run.vm04.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 651 kB/s | 84 kB 00:00 2026-03-10T12:54:12.436 INFO:teuthology.orchestra.run.vm04.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.1 MB/s | 150 kB 00:00 2026-03-10T12:54:12.442 INFO:teuthology.orchestra.run.vm09.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.2 MB/s | 150 kB 00:00 2026-03-10T12:54:12.508 INFO:teuthology.orchestra.run.vm09.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.8 MB/s | 1.2 MB 00:00 2026-03-10T12:54:12.564 INFO:teuthology.orchestra.run.vm09.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 4.5 MB/s | 837 kB 00:00 2026-03-10T12:54:12.564 INFO:teuthology.orchestra.run.vm04.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 2.2 MB/s | 1.2 MB 00:00 2026-03-10T12:54:12.564 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:54:12.564 INFO:teuthology.orchestra.run.vm09.stdout:Total 2.5 MB/s | 2.3 MB 00:00 2026-03-10T12:54:12.626 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T12:54:12.632 INFO:teuthology.orchestra.run.vm04.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 4.1 MB/s | 837 kB 00:00 2026-03-10T12:54:12.632 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:54:12.632 INFO:teuthology.orchestra.run.vm04.stdout:Total 2.2 MB/s | 2.3 MB 00:01 2026-03-10T12:54:12.633 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T12:54:12.633 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T12:54:12.682 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T12:54:12.683 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T12:54:12.695 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T12:54:12.703 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T12:54:12.703 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T12:54:12.754 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T12:54:12.755 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T12:54:12.777 INFO:teuthology.orchestra.run.vm06.stdout:(6/6): nvme-cli-2.16-1.el9.x86_64.rpm 1.3 MB/s | 1.2 MB 00:00 2026-03-10T12:54:12.777 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:54:12.777 INFO:teuthology.orchestra.run.vm06.stdout:Total 2.3 MB/s | 2.3 MB 00:01 2026-03-10T12:54:12.828 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T12:54:12.839 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T12:54:12.850 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-10T12:54:12.855 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T12:54:12.858 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:12.862 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T12:54:12.862 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T12:54:12.865 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:12.867 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:12.921 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T12:54:12.922 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T12:54:12.922 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T12:54:12.932 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T12:54:12.943 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-10T12:54:12.951 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:12.959 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:12.961 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:13.011 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:13.017 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.113 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T12:54:13.126 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:13.130 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-10T12:54:13.131 INFO:teuthology.orchestra.run.vm04.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.149 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-10T12:54:13.160 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:13.170 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:13.174 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:13.375 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-10T12:54:13.379 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.395 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.395 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T12:54:13.395 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:13.499 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.499 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T12:54:13.499 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:13.771 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-10T12:54:13.771 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T12:54:13.772 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:13.973 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T12:54:13.974 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T12:54:13.974 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:13.974 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:13.975 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-10T12:54:14.058 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:14.059 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T12:54:14.072 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T12:54:14.072 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T12:54:14.072 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:14.072 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:14.072 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-10T12:54:14.135 DEBUG:teuthology.parallel:result is None 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:14.181 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T12:54:14.248 DEBUG:teuthology.parallel:result is None 2026-03-10T12:54:14.409 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-10T12:54:14.409 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-10T12:54:14.409 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-10T12:54:14.409 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-10T12:54:14.409 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:14.509 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T12:54:14.582 DEBUG:teuthology.parallel:result is None 2026-03-10T12:54:14.582 INFO:teuthology.run_tasks:Running task install... 2026-03-10T12:54:14.584 DEBUG:teuthology.task.install:project ceph 2026-03-10T12:54:14.584 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T12:54:14.584 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-10T12:54:14.584 INFO:teuthology.task.install:Using flavor: default 2026-03-10T12:54:14.586 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-10T12:54:14.586 INFO:teuthology.task.install:extra packages: [] 2026-03-10T12:54:14.586 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T12:54:14.587 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:54:14.587 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T12:54:14.587 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:54:14.587 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-10T12:54:14.588 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:54:15.273 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T12:54:15.273 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T12:54:15.358 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T12:54:15.358 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T12:54:15.386 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-10T12:54:15.386 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-10T12:54:15.807 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T12:54:15.807 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:54:15.807 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T12:54:15.848 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T12:54:15.848 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:54:15.848 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T12:54:15.848 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T12:54:15.848 DEBUG:teuthology.orchestra.run.vm04:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T12:54:15.883 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-10T12:54:15.883 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:54:15.883 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-10T12:54:15.892 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T12:54:15.892 DEBUG:teuthology.orchestra.run.vm06:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T12:54:15.919 DEBUG:teuthology.orchestra.run.vm04:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T12:54:15.921 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-10T12:54:15.921 DEBUG:teuthology.orchestra.run.vm09:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-10T12:54:15.978 DEBUG:teuthology.orchestra.run.vm06:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T12:54:16.003 DEBUG:teuthology.orchestra.run.vm09:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-10T12:54:16.003 DEBUG:teuthology.orchestra.run.vm04:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T12:54:16.030 DEBUG:teuthology.orchestra.run.vm06:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T12:54:16.038 INFO:teuthology.orchestra.run.vm04.stdout:check_obsoletes = 1 2026-03-10T12:54:16.040 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-10T12:54:16.073 INFO:teuthology.orchestra.run.vm06.stdout:check_obsoletes = 1 2026-03-10T12:54:16.075 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-10T12:54:16.085 DEBUG:teuthology.orchestra.run.vm09:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-10T12:54:16.156 INFO:teuthology.orchestra.run.vm09.stdout:check_obsoletes = 1 2026-03-10T12:54:16.158 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean all 2026-03-10T12:54:16.242 INFO:teuthology.orchestra.run.vm04.stdout:41 files removed 2026-03-10T12:54:16.255 INFO:teuthology.orchestra.run.vm06.stdout:41 files removed 2026-03-10T12:54:16.282 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T12:54:16.294 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T12:54:16.351 INFO:teuthology.orchestra.run.vm09.stdout:41 files removed 2026-03-10T12:54:16.376 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-10T12:54:17.636 INFO:teuthology.orchestra.run.vm04.stdout:ceph packages for x86_64 71 kB/s | 84 kB 00:01 2026-03-10T12:54:17.696 INFO:teuthology.orchestra.run.vm06.stdout:ceph packages for x86_64 71 kB/s | 84 kB 00:01 2026-03-10T12:54:17.701 INFO:teuthology.orchestra.run.vm09.stdout:ceph packages for x86_64 73 kB/s | 84 kB 00:01 2026-03-10T12:54:18.629 INFO:teuthology.orchestra.run.vm04.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T12:54:18.670 INFO:teuthology.orchestra.run.vm09.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T12:54:18.681 INFO:teuthology.orchestra.run.vm06.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-10T12:54:19.593 INFO:teuthology.orchestra.run.vm04.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-10T12:54:19.625 INFO:teuthology.orchestra.run.vm09.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-10T12:54:19.632 INFO:teuthology.orchestra.run.vm06.stdout:ceph source packages 2.1 kB/s | 1.9 kB 00:00 2026-03-10T12:54:20.275 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - BaseOS 14 MB/s | 8.9 MB 00:00 2026-03-10T12:54:20.455 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - BaseOS 11 MB/s | 8.9 MB 00:00 2026-03-10T12:54:21.113 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - BaseOS 5.9 MB/s | 8.9 MB 00:01 2026-03-10T12:54:22.458 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - AppStream 20 MB/s | 27 MB 00:01 2026-03-10T12:54:23.393 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - AppStream 12 MB/s | 27 MB 00:02 2026-03-10T12:54:23.788 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - AppStream 13 MB/s | 27 MB 00:02 2026-03-10T12:54:26.738 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - CRB 7.6 MB/s | 8.0 MB 00:01 2026-03-10T12:54:27.039 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - CRB 7.9 MB/s | 8.0 MB 00:01 2026-03-10T12:54:27.844 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - CRB 5.8 MB/s | 8.0 MB 00:01 2026-03-10T12:54:28.280 INFO:teuthology.orchestra.run.vm06.stdout:CentOS Stream 9 - Extras packages 29 kB/s | 20 kB 00:00 2026-03-10T12:54:28.685 INFO:teuthology.orchestra.run.vm09.stdout:CentOS Stream 9 - Extras packages 26 kB/s | 20 kB 00:00 2026-03-10T12:54:28.727 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 55 MB/s | 20 MB 00:00 2026-03-10T12:54:29.085 INFO:teuthology.orchestra.run.vm04.stdout:CentOS Stream 9 - Extras packages 53 kB/s | 20 kB 00:00 2026-03-10T12:54:29.139 INFO:teuthology.orchestra.run.vm09.stdout:Extra Packages for Enterprise Linux 54 MB/s | 20 MB 00:00 2026-03-10T12:54:29.960 INFO:teuthology.orchestra.run.vm04.stdout:Extra Packages for Enterprise Linux 25 MB/s | 20 MB 00:00 2026-03-10T12:54:33.335 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-03-10T12:54:33.550 INFO:teuthology.orchestra.run.vm09.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-10T12:54:34.515 INFO:teuthology.orchestra.run.vm04.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-10T12:54:34.661 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:34.662 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:34.665 INFO:teuthology.orchestra.run.vm06.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T12:54:34.666 INFO:teuthology.orchestra.run.vm06.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T12:54:34.693 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T12:54:34.697 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T12:54:34.698 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T12:54:34.699 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:====================================================================================== 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Install 134 Packages 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 210 M 2026-03-10T12:54:34.700 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T12:54:34.898 INFO:teuthology.orchestra.run.vm09.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:34.899 INFO:teuthology.orchestra.run.vm09.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:34.903 INFO:teuthology.orchestra.run.vm09.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T12:54:34.903 INFO:teuthology.orchestra.run.vm09.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T12:54:34.932 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout:Upgrading: 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T12:54:34.937 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T12:54:34.938 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Installing weak dependencies: 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:====================================================================================== 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Install 134 Packages 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Upgrade 2 Packages 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 210 M 2026-03-10T12:54:34.939 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T12:54:35.889 INFO:teuthology.orchestra.run.vm04.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:35.890 INFO:teuthology.orchestra.run.vm04.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-10T12:54:35.893 INFO:teuthology.orchestra.run.vm04.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-10T12:54:35.894 INFO:teuthology.orchestra.run.vm04.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-10T12:54:35.921 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T12:54:35.925 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout:Installing: 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout:Upgrading: 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout:Installing dependencies: 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-10T12:54:35.926 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-10T12:54:35.927 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Installing weak dependencies: 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:====================================================================================== 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Install 134 Packages 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Upgrade 2 Packages 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Total download size: 210 M 2026-03-10T12:54:35.928 INFO:teuthology.orchestra.run.vm04.stdout:Downloading Packages: 2026-03-10T12:54:36.343 INFO:teuthology.orchestra.run.vm06.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T12:54:37.068 INFO:teuthology.orchestra.run.vm09.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T12:54:37.536 INFO:teuthology.orchestra.run.vm04.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-10T12:54:37.733 INFO:teuthology.orchestra.run.vm06.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 848 kB/s | 1.2 MB 00:01 2026-03-10T12:54:37.967 INFO:teuthology.orchestra.run.vm06.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 625 kB/s | 145 kB 00:00 2026-03-10T12:54:38.486 INFO:teuthology.orchestra.run.vm09.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 831 kB/s | 1.2 MB 00:01 2026-03-10T12:54:38.717 INFO:teuthology.orchestra.run.vm09.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 630 kB/s | 145 kB 00:00 2026-03-10T12:54:38.915 INFO:teuthology.orchestra.run.vm04.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 855 kB/s | 1.2 MB 00:01 2026-03-10T12:54:39.140 INFO:teuthology.orchestra.run.vm04.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 646 kB/s | 145 kB 00:00 2026-03-10T12:54:39.365 INFO:teuthology.orchestra.run.vm06.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 1.7 MB/s | 2.4 MB 00:01 2026-03-10T12:54:39.764 INFO:teuthology.orchestra.run.vm06.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 5.5 MB 00:03 2026-03-10T12:54:40.207 INFO:teuthology.orchestra.run.vm06.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 1.3 MB/s | 1.1 MB 00:00 2026-03-10T12:54:40.948 INFO:teuthology.orchestra.run.vm04.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 1.3 MB/s | 2.4 MB 00:01 2026-03-10T12:54:41.163 INFO:teuthology.orchestra.run.vm09.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 1.0 MB/s | 2.4 MB 00:02 2026-03-10T12:54:41.732 INFO:teuthology.orchestra.run.vm04.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 1.1 MB 00:00 2026-03-10T12:54:42.054 INFO:teuthology.orchestra.run.vm09.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.0 MB/s | 5.5 MB 00:05 2026-03-10T12:54:42.311 INFO:teuthology.orchestra.run.vm09.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 959 kB/s | 1.1 MB 00:01 2026-03-10T12:54:42.511 INFO:teuthology.orchestra.run.vm04.stdout:(6/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.0 MB/s | 5.5 MB 00:05 2026-03-10T12:54:42.820 INFO:teuthology.orchestra.run.vm06.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 1.5 MB/s | 4.7 MB 00:03 2026-03-10T12:54:44.636 INFO:teuthology.orchestra.run.vm04.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 1.6 MB/s | 4.7 MB 00:02 2026-03-10T12:54:45.909 INFO:teuthology.orchestra.run.vm09.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 1.2 MB/s | 4.7 MB 00:03 2026-03-10T12:54:48.459 INFO:teuthology.orchestra.run.vm06.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 1.9 MB/s | 11 MB 00:05 2026-03-10T12:54:48.572 INFO:teuthology.orchestra.run.vm06.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 221 kB/s | 25 kB 00:00 2026-03-10T12:54:49.900 INFO:teuthology.orchestra.run.vm04.stdout:(8/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 1.7 MB/s | 22 MB 00:12 2026-03-10T12:54:50.020 INFO:teuthology.orchestra.run.vm04.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 209 kB/s | 25 kB 00:00 2026-03-10T12:54:50.485 INFO:teuthology.orchestra.run.vm04.stdout:(10/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 1.8 MB/s | 11 MB 00:05 2026-03-10T12:54:50.517 INFO:teuthology.orchestra.run.vm06.stdout:(10/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 17 MB 00:10 2026-03-10T12:54:50.598 INFO:teuthology.orchestra.run.vm04.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 298 kB/s | 34 kB 00:00 2026-03-10T12:54:50.634 INFO:teuthology.orchestra.run.vm06.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 287 kB/s | 34 kB 00:00 2026-03-10T12:54:51.049 INFO:teuthology.orchestra.run.vm04.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.2 MB/s | 1.0 MB 00:00 2026-03-10T12:54:51.106 INFO:teuthology.orchestra.run.vm06.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.1 MB/s | 1.0 MB 00:00 2026-03-10T12:54:51.166 INFO:teuthology.orchestra.run.vm04.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-10T12:54:51.224 INFO:teuthology.orchestra.run.vm06.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-10T12:54:51.279 INFO:teuthology.orchestra.run.vm04.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T12:54:51.342 INFO:teuthology.orchestra.run.vm06.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T12:54:51.507 INFO:teuthology.orchestra.run.vm04.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 2.2 MB/s | 503 kB 00:00 2026-03-10T12:54:51.577 INFO:teuthology.orchestra.run.vm06.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 2.1 MB/s | 503 kB 00:00 2026-03-10T12:54:51.851 INFO:teuthology.orchestra.run.vm09.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 1.8 MB/s | 11 MB 00:05 2026-03-10T12:54:52.053 INFO:teuthology.orchestra.run.vm09.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 124 kB/s | 25 kB 00:00 2026-03-10T12:54:52.201 INFO:teuthology.orchestra.run.vm06.stdout:(16/136): ceph-common-19.2.3-678.ge911bdeb.el9. 1.3 MB/s | 22 MB 00:16 2026-03-10T12:54:52.261 INFO:teuthology.orchestra.run.vm09.stdout:(10/136): ceph-common-19.2.3-678.ge911bdeb.el9. 1.4 MB/s | 22 MB 00:15 2026-03-10T12:54:52.321 INFO:teuthology.orchestra.run.vm06.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 376 kB/s | 45 kB 00:00 2026-03-10T12:54:52.379 INFO:teuthology.orchestra.run.vm09.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 287 kB/s | 34 kB 00:00 2026-03-10T12:54:52.462 INFO:teuthology.orchestra.run.vm09.stdout:(12/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 17 MB 00:10 2026-03-10T12:54:52.558 INFO:teuthology.orchestra.run.vm06.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 600 kB/s | 142 kB 00:00 2026-03-10T12:54:52.579 INFO:teuthology.orchestra.run.vm09.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.4 MB/s | 163 kB 00:00 2026-03-10T12:54:52.679 INFO:teuthology.orchestra.run.vm06.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-10T12:54:52.696 INFO:teuthology.orchestra.run.vm09.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 1.1 MB/s | 127 kB 00:00 2026-03-10T12:54:52.780 INFO:teuthology.orchestra.run.vm04.stdout:(16/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 17 MB 00:10 2026-03-10T12:54:52.846 INFO:teuthology.orchestra.run.vm09.stdout:(15/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 2.1 MB/s | 1.0 MB 00:00 2026-03-10T12:54:52.899 INFO:teuthology.orchestra.run.vm04.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 380 kB/s | 45 kB 00:00 2026-03-10T12:54:52.917 INFO:teuthology.orchestra.run.vm06.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 1.3 MB/s | 323 kB 00:00 2026-03-10T12:54:52.928 INFO:teuthology.orchestra.run.vm09.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 2.1 MB/s | 503 kB 00:00 2026-03-10T12:54:53.019 INFO:teuthology.orchestra.run.vm04.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-10T12:54:53.044 INFO:teuthology.orchestra.run.vm09.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 389 kB/s | 45 kB 00:00 2026-03-10T12:54:53.139 INFO:teuthology.orchestra.run.vm04.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-10T12:54:53.139 INFO:teuthology.orchestra.run.vm06.stdout:(21/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 3.5 MB/s | 5.4 MB 00:01 2026-03-10T12:54:53.155 INFO:teuthology.orchestra.run.vm06.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 1.2 MB/s | 303 kB 00:00 2026-03-10T12:54:53.161 INFO:teuthology.orchestra.run.vm09.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.2 MB/s | 142 kB 00:00 2026-03-10T12:54:53.273 INFO:teuthology.orchestra.run.vm06.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 743 kB/s | 100 kB 00:00 2026-03-10T12:54:53.274 INFO:teuthology.orchestra.run.vm06.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 711 kB/s | 85 kB 00:00 2026-03-10T12:54:53.278 INFO:teuthology.orchestra.run.vm09.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.4 MB/s | 165 kB 00:00 2026-03-10T12:54:53.343 INFO:teuthology.orchestra.run.vm04.stdout:(20/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.9 MB/s | 5.4 MB 00:01 2026-03-10T12:54:53.376 INFO:teuthology.orchestra.run.vm04.stdout:(21/136): python3-rados-19.2.3-678.ge911bdeb.el 1.3 MB/s | 323 kB 00:00 2026-03-10T12:54:53.396 INFO:teuthology.orchestra.run.vm06.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-10T12:54:53.401 INFO:teuthology.orchestra.run.vm09.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.6 MB/s | 323 kB 00:00 2026-03-10T12:54:53.458 INFO:teuthology.orchestra.run.vm04.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.6 MB/s | 303 kB 00:00 2026-03-10T12:54:53.495 INFO:teuthology.orchestra.run.vm04.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 837 kB/s | 100 kB 00:00 2026-03-10T12:54:53.515 INFO:teuthology.orchestra.run.vm06.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 261 kB/s | 31 kB 00:00 2026-03-10T12:54:53.520 INFO:teuthology.orchestra.run.vm09.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.5 MB/s | 303 kB 00:00 2026-03-10T12:54:53.571 INFO:teuthology.orchestra.run.vm04.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 753 kB/s | 85 kB 00:00 2026-03-10T12:54:53.636 INFO:teuthology.orchestra.run.vm06.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.2 MB/s | 150 kB 00:00 2026-03-10T12:54:53.637 INFO:teuthology.orchestra.run.vm09.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 854 kB/s | 100 kB 00:00 2026-03-10T12:54:53.686 INFO:teuthology.orchestra.run.vm04.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.5 MB/s | 171 kB 00:00 2026-03-10T12:54:53.754 INFO:teuthology.orchestra.run.vm09.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 729 kB/s | 85 kB 00:00 2026-03-10T12:54:53.799 INFO:teuthology.orchestra.run.vm04.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 276 kB/s | 31 kB 00:00 2026-03-10T12:54:53.913 INFO:teuthology.orchestra.run.vm04.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 1.3 MB/s | 150 kB 00:00 2026-03-10T12:54:53.982 INFO:teuthology.orchestra.run.vm06.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 4.4 MB/s | 3.1 MB 00:00 2026-03-10T12:54:54.570 INFO:teuthology.orchestra.run.vm09.stdout:(24/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 3.8 MB/s | 3.1 MB 00:00 2026-03-10T12:54:54.688 INFO:teuthology.orchestra.run.vm09.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-10T12:54:54.804 INFO:teuthology.orchestra.run.vm09.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 269 kB/s | 31 kB 00:00 2026-03-10T12:54:54.836 INFO:teuthology.orchestra.run.vm09.stdout:(27/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.7 MB/s | 5.4 MB 00:01 2026-03-10T12:54:54.933 INFO:teuthology.orchestra.run.vm04.stdout:(28/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 3.7 MB/s | 3.8 MB 00:01 2026-03-10T12:54:55.200 INFO:teuthology.orchestra.run.vm06.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 2.4 MB/s | 3.8 MB 00:01 2026-03-10T12:54:55.440 INFO:teuthology.orchestra.run.vm06.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.0 MB/s | 253 kB 00:00 2026-03-10T12:54:55.555 INFO:teuthology.orchestra.run.vm04.stdout:(29/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 1.5 MB/s | 3.1 MB 00:02 2026-03-10T12:54:55.562 INFO:teuthology.orchestra.run.vm06.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 404 kB/s | 49 kB 00:00 2026-03-10T12:54:55.681 INFO:teuthology.orchestra.run.vm06.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 141 kB/s | 17 kB 00:00 2026-03-10T12:54:55.795 INFO:teuthology.orchestra.run.vm04.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.0 MB/s | 253 kB 00:00 2026-03-10T12:54:55.883 INFO:teuthology.orchestra.run.vm06.stdout:(33/136): ceph-mgr-diskprediction-local-19.2.3- 3.9 MB/s | 7.4 MB 00:01 2026-03-10T12:54:55.907 INFO:teuthology.orchestra.run.vm09.stdout:(28/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 136 kB/s | 150 kB 00:01 2026-03-10T12:54:55.919 INFO:teuthology.orchestra.run.vm04.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 397 kB/s | 49 kB 00:00 2026-03-10T12:54:56.042 INFO:teuthology.orchestra.run.vm04.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 137 kB/s | 17 kB 00:00 2026-03-10T12:54:56.280 INFO:teuthology.orchestra.run.vm04.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 1.2 MB/s | 299 kB 00:00 2026-03-10T12:54:56.674 INFO:teuthology.orchestra.run.vm06.stdout:(34/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 301 kB/s | 299 kB 00:00 2026-03-10T12:54:56.702 INFO:teuthology.orchestra.run.vm06.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 940 kB/s | 769 kB 00:00 2026-03-10T12:54:56.759 INFO:teuthology.orchestra.run.vm09.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 2.0 MB/s | 3.8 MB 00:01 2026-03-10T12:54:56.844 INFO:teuthology.orchestra.run.vm06.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 284 kB/s | 40 kB 00:00 2026-03-10T12:54:56.870 INFO:teuthology.orchestra.run.vm04.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 1.3 MB/s | 769 kB 00:00 2026-03-10T12:54:56.940 INFO:teuthology.orchestra.run.vm06.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 750 kB/s | 72 kB 00:00 2026-03-10T12:54:56.957 INFO:teuthology.orchestra.run.vm06.stdout:(38/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.2 MB/s | 351 kB 00:00 2026-03-10T12:54:56.994 INFO:teuthology.orchestra.run.vm09.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.1 MB/s | 253 kB 00:00 2026-03-10T12:54:57.009 INFO:teuthology.orchestra.run.vm06.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 3.5 MB/s | 184 kB 00:00 2026-03-10T12:54:57.057 INFO:teuthology.orchestra.run.vm06.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 692 kB/s | 33 kB 00:00 2026-03-10T12:54:57.088 INFO:teuthology.orchestra.run.vm06.stdout:(41/136): libgfortran-11.5.0-14.el9.x86_64.rpm 5.3 MB/s | 794 kB 00:00 2026-03-10T12:54:57.107 INFO:teuthology.orchestra.run.vm06.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 1.8 MB/s | 93 kB 00:00 2026-03-10T12:54:57.112 INFO:teuthology.orchestra.run.vm09.stdout:(31/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 418 kB/s | 49 kB 00:00 2026-03-10T12:54:57.140 INFO:teuthology.orchestra.run.vm06.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 4.8 MB/s | 253 kB 00:00 2026-03-10T12:54:57.236 INFO:teuthology.orchestra.run.vm09.stdout:(32/136): ceph-prometheus-alerts-19.2.3-678.ge9 134 kB/s | 17 kB 00:00 2026-03-10T12:54:57.237 INFO:teuthology.orchestra.run.vm06.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 1.1 MB/s | 106 kB 00:00 2026-03-10T12:54:57.287 INFO:teuthology.orchestra.run.vm06.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 2.6 MB/s | 135 kB 00:00 2026-03-10T12:54:57.392 INFO:teuthology.orchestra.run.vm06.stdout:(46/136): python3-cryptography-36.0.1-5.el9.x86 4.4 MB/s | 1.2 MB 00:00 2026-03-10T12:54:57.392 INFO:teuthology.orchestra.run.vm04.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 673 kB/s | 351 kB 00:00 2026-03-10T12:54:57.394 INFO:teuthology.orchestra.run.vm06.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 1.2 MB/s | 126 kB 00:00 2026-03-10T12:54:57.444 INFO:teuthology.orchestra.run.vm06.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 4.1 MB/s | 218 kB 00:00 2026-03-10T12:54:57.445 INFO:teuthology.orchestra.run.vm06.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 3.4 MB/s | 182 kB 00:00 2026-03-10T12:54:57.471 INFO:teuthology.orchestra.run.vm09.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 1.2 MB/s | 299 kB 00:00 2026-03-10T12:54:57.533 INFO:teuthology.orchestra.run.vm06.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 2.9 MB/s | 266 kB 00:00 2026-03-10T12:54:57.533 INFO:teuthology.orchestra.run.vm04.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 287 kB/s | 40 kB 00:00 2026-03-10T12:54:57.697 INFO:teuthology.orchestra.run.vm06.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 184 kB/s | 30 kB 00:00 2026-03-10T12:54:57.697 INFO:teuthology.orchestra.run.vm04.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 440 kB/s | 72 kB 00:00 2026-03-10T12:54:57.718 INFO:teuthology.orchestra.run.vm06.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 383 kB/s | 104 kB 00:00 2026-03-10T12:54:57.824 INFO:teuthology.orchestra.run.vm06.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 139 kB/s | 15 kB 00:00 2026-03-10T12:54:57.864 INFO:teuthology.orchestra.run.vm06.stdout:(54/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 18 MB/s | 3.0 MB 00:00 2026-03-10T12:54:57.893 INFO:teuthology.orchestra.run.vm06.stdout:(55/136): libnbd-1.20.3-4.el9.x86_64.rpm 2.3 MB/s | 164 kB 00:00 2026-03-10T12:54:57.900 INFO:teuthology.orchestra.run.vm04.stdout:(38/136): libgfortran-11.5.0-14.el9.x86_64.rpm 3.8 MB/s | 794 kB 00:00 2026-03-10T12:54:57.925 INFO:teuthology.orchestra.run.vm06.stdout:(56/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.6 MB/s | 160 kB 00:00 2026-03-10T12:54:57.947 INFO:teuthology.orchestra.run.vm06.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 847 kB/s | 45 kB 00:00 2026-03-10T12:54:58.026 INFO:teuthology.orchestra.run.vm06.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 3.0 MB/s | 246 kB 00:00 2026-03-10T12:54:58.027 INFO:teuthology.orchestra.run.vm04.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 1.4 MB/s | 184 kB 00:00 2026-03-10T12:54:58.027 INFO:teuthology.orchestra.run.vm09.stdout:(34/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 1.4 MB/s | 769 kB 00:00 2026-03-10T12:54:58.029 INFO:teuthology.orchestra.run.vm06.stdout:(59/136): librdkafka-1.6.1-102.el9.x86_64.rpm 6.2 MB/s | 662 kB 00:00 2026-03-10T12:54:58.086 INFO:teuthology.orchestra.run.vm06.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 3.9 MB/s | 233 kB 00:00 2026-03-10T12:54:58.089 INFO:teuthology.orchestra.run.vm06.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 4.8 MB/s | 292 kB 00:00 2026-03-10T12:54:58.142 INFO:teuthology.orchestra.run.vm04.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 289 kB/s | 33 kB 00:00 2026-03-10T12:54:58.145 INFO:teuthology.orchestra.run.vm06.stdout:(62/136): openblas-0.3.29-1.el9.x86_64.rpm 754 kB/s | 42 kB 00:00 2026-03-10T12:54:58.153 INFO:teuthology.orchestra.run.vm06.stdout:(63/136): lua-5.4.4-4.el9.x86_64.rpm 2.8 MB/s | 188 kB 00:00 2026-03-10T12:54:58.229 INFO:teuthology.orchestra.run.vm04.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 1.0 MB/s | 93 kB 00:00 2026-03-10T12:54:58.273 INFO:teuthology.orchestra.run.vm04.stdout:(42/136): ceph-mgr-diskprediction-local-19.2.3- 2.2 MB/s | 7.4 MB 00:03 2026-03-10T12:54:58.312 INFO:teuthology.orchestra.run.vm06.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 6.3 MB/s | 1.0 MB 00:00 2026-03-10T12:54:58.328 INFO:teuthology.orchestra.run.vm06.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 29 MB/s | 5.3 MB 00:00 2026-03-10T12:54:58.328 INFO:teuthology.orchestra.run.vm04.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 2.5 MB/s | 253 kB 00:00 2026-03-10T12:54:58.402 INFO:teuthology.orchestra.run.vm06.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 3.2 MB/s | 244 kB 00:00 2026-03-10T12:54:58.412 INFO:teuthology.orchestra.run.vm04.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 1.2 MB/s | 106 kB 00:00 2026-03-10T12:54:58.415 INFO:teuthology.orchestra.run.vm09.stdout:(35/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 908 kB/s | 351 kB 00:00 2026-03-10T12:54:58.472 INFO:teuthology.orchestra.run.vm06.stdout:(67/136): python3-babel-2.9.1-2.el9.noarch.rpm 38 MB/s | 6.0 MB 00:00 2026-03-10T12:54:58.474 INFO:teuthology.orchestra.run.vm06.stdout:(68/136): python3-jinja2-2.11.3-8.el9.noarch.rp 3.4 MB/s | 249 kB 00:00 2026-03-10T12:54:58.502 INFO:teuthology.orchestra.run.vm09.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 465 kB/s | 40 kB 00:00 2026-03-10T12:54:58.511 INFO:teuthology.orchestra.run.vm04.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 1.3 MB/s | 135 kB 00:00 2026-03-10T12:54:58.529 INFO:teuthology.orchestra.run.vm06.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 843 kB/s | 48 kB 00:00 2026-03-10T12:54:58.537 INFO:teuthology.orchestra.run.vm06.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 2.8 MB/s | 177 kB 00:00 2026-03-10T12:54:58.576 INFO:teuthology.orchestra.run.vm09.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 981 kB/s | 72 kB 00:00 2026-03-10T12:54:58.588 INFO:teuthology.orchestra.run.vm06.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.9 MB/s | 172 kB 00:00 2026-03-10T12:54:58.595 INFO:teuthology.orchestra.run.vm06.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 595 kB/s | 35 kB 00:00 2026-03-10T12:54:58.600 INFO:teuthology.orchestra.run.vm04.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 1.4 MB/s | 126 kB 00:00 2026-03-10T12:54:58.695 INFO:teuthology.orchestra.run.vm04.stdout:(47/136): python3-urllib3-1.26.5-7.el9.noarch.r 2.3 MB/s | 218 kB 00:00 2026-03-10T12:54:58.701 INFO:teuthology.orchestra.run.vm06.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 4.1 MB/s | 442 kB 00:00 2026-03-10T12:54:58.735 INFO:teuthology.orchestra.run.vm06.stdout:(74/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 42 MB/s | 6.1 MB 00:00 2026-03-10T12:54:58.747 INFO:teuthology.orchestra.run.vm09.stdout:(38/136): ceph-mgr-diskprediction-local-19.2.3- 2.6 MB/s | 7.4 MB 00:02 2026-03-10T12:54:58.751 INFO:teuthology.orchestra.run.vm09.stdout:(39/136): libgfortran-11.5.0-14.el9.x86_64.rpm 4.4 MB/s | 794 kB 00:00 2026-03-10T12:54:58.753 INFO:teuthology.orchestra.run.vm06.stdout:(75/136): python3-packaging-20.9-5.el9.noarch.r 1.5 MB/s | 77 kB 00:00 2026-03-10T12:54:58.788 INFO:teuthology.orchestra.run.vm04.stdout:(48/136): unzip-6.0-59.el9.x86_64.rpm 1.9 MB/s | 182 kB 00:00 2026-03-10T12:54:58.794 INFO:teuthology.orchestra.run.vm06.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 4.5 MB/s | 267 kB 00:00 2026-03-10T12:54:58.810 INFO:teuthology.orchestra.run.vm06.stdout:(77/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 2.8 MB/s | 157 kB 00:00 2026-03-10T12:54:58.811 INFO:teuthology.orchestra.run.vm09.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 548 kB/s | 33 kB 00:00 2026-03-10T12:54:58.854 INFO:teuthology.orchestra.run.vm06.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 4.5 MB/s | 277 kB 00:00 2026-03-10T12:54:58.867 INFO:teuthology.orchestra.run.vm06.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 936 kB/s | 54 kB 00:00 2026-03-10T12:54:58.907 INFO:teuthology.orchestra.run.vm04.stdout:(49/136): zip-3.0-35.el9.x86_64.rpm 2.2 MB/s | 266 kB 00:00 2026-03-10T12:54:58.910 INFO:teuthology.orchestra.run.vm09.stdout:(41/136): pciutils-3.7.0-7.el9.x86_64.rpm 951 kB/s | 93 kB 00:00 2026-03-10T12:54:58.932 INFO:teuthology.orchestra.run.vm04.stdout:(50/136): python3-cryptography-36.0.1-5.el9.x86 1.9 MB/s | 1.2 MB 00:00 2026-03-10T12:54:58.944 INFO:teuthology.orchestra.run.vm06.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 542 kB/s | 42 kB 00:00 2026-03-10T12:54:58.993 INFO:teuthology.orchestra.run.vm09.stdout:(42/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.0 MB/s | 253 kB 00:00 2026-03-10T12:54:59.077 INFO:teuthology.orchestra.run.vm09.stdout:(43/136): libquadmath-11.5.0-14.el9.x86_64.rpm 561 kB/s | 184 kB 00:00 2026-03-10T12:54:59.105 INFO:teuthology.orchestra.run.vm06.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 1.5 MB/s | 240 kB 00:00 2026-03-10T12:54:59.125 INFO:teuthology.orchestra.run.vm04.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 154 kB/s | 30 kB 00:00 2026-03-10T12:54:59.181 INFO:teuthology.orchestra.run.vm09.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 1.0 MB/s | 106 kB 00:00 2026-03-10T12:54:59.183 INFO:teuthology.orchestra.run.vm06.stdout:(82/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 59 MB/s | 19 MB 00:00 2026-03-10T12:54:59.184 INFO:teuthology.orchestra.run.vm06.stdout:(83/136): qatlib-service-25.08.0-2.el9.x86_64.r 467 kB/s | 37 kB 00:00 2026-03-10T12:54:59.225 INFO:teuthology.orchestra.run.vm04.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 327 kB/s | 104 kB 00:00 2026-03-10T12:54:59.240 INFO:teuthology.orchestra.run.vm06.stdout:(84/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.1 MB/s | 66 kB 00:00 2026-03-10T12:54:59.251 INFO:teuthology.orchestra.run.vm06.stdout:(85/136): socat-1.7.4.1-8.el9.x86_64.rpm 4.5 MB/s | 303 kB 00:00 2026-03-10T12:54:59.284 INFO:teuthology.orchestra.run.vm09.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 1.3 MB/s | 135 kB 00:00 2026-03-10T12:54:59.291 INFO:teuthology.orchestra.run.vm04.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 228 kB/s | 15 kB 00:00 2026-03-10T12:54:59.307 INFO:teuthology.orchestra.run.vm06.stdout:(86/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 961 kB/s | 64 kB 00:00 2026-03-10T12:54:59.423 INFO:teuthology.orchestra.run.vm04.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.2 MB/s | 164 kB 00:00 2026-03-10T12:54:59.436 INFO:teuthology.orchestra.run.vm06.stdout:(87/136): lua-devel-5.4.4-4.el9.x86_64.rpm 121 kB/s | 22 kB 00:00 2026-03-10T12:54:59.459 INFO:teuthology.orchestra.run.vm09.stdout:(46/136): python3-requests-2.25.1-10.el9.noarch 720 kB/s | 126 kB 00:00 2026-03-10T12:54:59.462 INFO:teuthology.orchestra.run.vm06.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 20 MB/s | 551 kB 00:00 2026-03-10T12:54:59.473 INFO:teuthology.orchestra.run.vm06.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 30 MB/s | 308 kB 00:00 2026-03-10T12:54:59.476 INFO:teuthology.orchestra.run.vm06.stdout:(90/136): protobuf-compiler-3.14.0-17.el9.x86_6 5.0 MB/s | 862 kB 00:00 2026-03-10T12:54:59.477 INFO:teuthology.orchestra.run.vm06.stdout:(91/136): grpc-data-1.46.7-10.el9.noarch.rpm 5.3 MB/s | 19 kB 00:00 2026-03-10T12:54:59.479 INFO:teuthology.orchestra.run.vm06.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 11 MB/s | 25 kB 00:00 2026-03-10T12:54:59.484 INFO:teuthology.orchestra.run.vm06.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 11 MB/s | 49 kB 00:00 2026-03-10T12:54:59.487 INFO:teuthology.orchestra.run.vm06.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 21 MB/s | 67 kB 00:00 2026-03-10T12:54:59.495 INFO:teuthology.orchestra.run.vm04.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 2.2 MB/s | 160 kB 00:00 2026-03-10T12:54:59.496 INFO:teuthology.orchestra.run.vm06.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 17 MB/s | 151 kB 00:00 2026-03-10T12:54:59.519 INFO:teuthology.orchestra.run.vm09.stdout:(47/136): python3-cryptography-36.0.1-5.el9.x86 2.4 MB/s | 1.2 MB 00:00 2026-03-10T12:54:59.533 INFO:teuthology.orchestra.run.vm06.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 23 MB/s | 838 kB 00:00 2026-03-10T12:54:59.533 INFO:teuthology.orchestra.run.vm09.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 2.9 MB/s | 218 kB 00:00 2026-03-10T12:54:59.552 INFO:teuthology.orchestra.run.vm06.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 28 MB/s | 548 kB 00:00 2026-03-10T12:54:59.555 INFO:teuthology.orchestra.run.vm06.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 11 MB/s | 29 kB 00:00 2026-03-10T12:54:59.559 INFO:teuthology.orchestra.run.vm06.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 18 MB/s | 60 kB 00:00 2026-03-10T12:54:59.574 INFO:teuthology.orchestra.run.vm06.stdout:(100/136): libarrow-9.0.0-15.el9.x86_64.rpm 45 MB/s | 4.4 MB 00:00 2026-03-10T12:54:59.574 INFO:teuthology.orchestra.run.vm04.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 571 kB/s | 45 kB 00:00 2026-03-10T12:54:59.575 INFO:teuthology.orchestra.run.vm06.stdout:(101/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 2.6 MB/s | 43 kB 00:00 2026-03-10T12:54:59.579 INFO:teuthology.orchestra.run.vm06.stdout:(102/136): python3-cachetools-4.2.4-1.el9.noarc 6.8 MB/s | 32 kB 00:00 2026-03-10T12:54:59.579 INFO:teuthology.orchestra.run.vm06.stdout:(103/136): python3-certifi-2023.05.07-4.el9.noa 3.4 MB/s | 14 kB 00:00 2026-03-10T12:54:59.588 INFO:teuthology.orchestra.run.vm09.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 2.6 MB/s | 182 kB 00:00 2026-03-10T12:54:59.589 INFO:teuthology.orchestra.run.vm06.stdout:(104/136): python3-cheroot-10.0.1-4.el9.noarch. 19 MB/s | 173 kB 00:00 2026-03-10T12:54:59.590 INFO:teuthology.orchestra.run.vm06.stdout:(105/136): python3-cherrypy-18.6.1-2.el9.noarch 33 MB/s | 358 kB 00:00 2026-03-10T12:54:59.597 INFO:teuthology.orchestra.run.vm06.stdout:(106/136): python3-google-auth-2.45.0-1.el9.noa 33 MB/s | 254 kB 00:00 2026-03-10T12:54:59.605 INFO:teuthology.orchestra.run.vm06.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 17 MB/s | 144 kB 00:00 2026-03-10T12:54:59.608 INFO:teuthology.orchestra.run.vm06.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 3.9 MB/s | 11 kB 00:00 2026-03-10T12:54:59.611 INFO:teuthology.orchestra.run.vm06.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 5.9 MB/s | 18 kB 00:00 2026-03-10T12:54:59.626 INFO:teuthology.orchestra.run.vm09.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 2.8 MB/s | 266 kB 00:00 2026-03-10T12:54:59.631 INFO:teuthology.orchestra.run.vm04.stdout:(57/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 5.9 MB/s | 3.0 MB 00:00 2026-03-10T12:54:59.632 INFO:teuthology.orchestra.run.vm06.stdout:(110/136): python3-grpcio-1.46.7-10.el9.x86_64. 50 MB/s | 2.0 MB 00:00 2026-03-10T12:54:59.632 INFO:teuthology.orchestra.run.vm06.stdout:(111/136): python3-jaraco-collections-3.0.0-8.e 1.1 MB/s | 23 kB 00:00 2026-03-10T12:54:59.634 INFO:teuthology.orchestra.run.vm06.stdout:(112/136): python3-jaraco-context-6.0.1-3.el9.n 8.7 MB/s | 20 kB 00:00 2026-03-10T12:54:59.637 INFO:teuthology.orchestra.run.vm06.stdout:(113/136): python3-jaraco-functools-3.5.0-2.el9 4.3 MB/s | 19 kB 00:00 2026-03-10T12:54:59.638 INFO:teuthology.orchestra.run.vm06.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 7.3 MB/s | 26 kB 00:00 2026-03-10T12:54:59.641 INFO:teuthology.orchestra.run.vm06.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 16 MB/s | 46 kB 00:00 2026-03-10T12:54:59.646 INFO:teuthology.orchestra.run.vm06.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 17 MB/s | 79 kB 00:00 2026-03-10T12:54:59.653 INFO:teuthology.orchestra.run.vm06.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 8.8 MB/s | 58 kB 00:00 2026-03-10T12:54:59.660 INFO:teuthology.orchestra.run.vm06.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 40 MB/s | 272 kB 00:00 2026-03-10T12:54:59.668 INFO:teuthology.orchestra.run.vm06.stdout:(119/136): python3-kubernetes-26.1.0-3.el9.noar 34 MB/s | 1.0 MB 00:00 2026-03-10T12:54:59.671 INFO:teuthology.orchestra.run.vm06.stdout:(120/136): python3-portend-3.1.0-2.el9.noarch.r 1.6 MB/s | 16 kB 00:00 2026-03-10T12:54:59.675 INFO:teuthology.orchestra.run.vm06.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 25 MB/s | 90 kB 00:00 2026-03-10T12:54:59.676 INFO:teuthology.orchestra.run.vm06.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 5.7 MB/s | 31 kB 00:00 2026-03-10T12:54:59.680 INFO:teuthology.orchestra.run.vm06.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 19 MB/s | 59 kB 00:00 2026-03-10T12:54:59.681 INFO:teuthology.orchestra.run.vm06.stdout:(124/136): python3-routes-2.5.1-5.el9.noarch.rp 31 MB/s | 188 kB 00:00 2026-03-10T12:54:59.686 INFO:teuthology.orchestra.run.vm06.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 6.2 MB/s | 36 kB 00:00 2026-03-10T12:54:59.687 INFO:teuthology.orchestra.run.vm06.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 14 MB/s | 86 kB 00:00 2026-03-10T12:54:59.705 INFO:teuthology.orchestra.run.vm04.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 3.3 MB/s | 246 kB 00:00 2026-03-10T12:54:59.705 INFO:teuthology.orchestra.run.vm06.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 5.1 MB/s | 90 kB 00:00 2026-03-10T12:54:59.708 INFO:teuthology.orchestra.run.vm06.stdout:(128/136): python3-webob-1.8.8-2.el9.noarch.rpm 10 MB/s | 230 kB 00:00 2026-03-10T12:54:59.709 INFO:teuthology.orchestra.run.vm04.stdout:(59/136): librdkafka-1.6.1-102.el9.x86_64.rpm 4.8 MB/s | 662 kB 00:00 2026-03-10T12:54:59.711 INFO:teuthology.orchestra.run.vm06.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 8.5 MB/s | 22 kB 00:00 2026-03-10T12:54:59.716 INFO:teuthology.orchestra.run.vm06.stdout:(130/136): python3-werkzeug-2.0.3-3.el9.1.noarc 38 MB/s | 427 kB 00:00 2026-03-10T12:54:59.718 INFO:teuthology.orchestra.run.vm06.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 2.9 MB/s | 20 kB 00:00 2026-03-10T12:54:59.718 INFO:teuthology.orchestra.run.vm09.stdout:(51/136): flexiblas-3.0.4-9.el9.x86_64.rpm 322 kB/s | 30 kB 00:00 2026-03-10T12:54:59.725 INFO:teuthology.orchestra.run.vm06.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 32 MB/s | 191 kB 00:00 2026-03-10T12:54:59.739 INFO:teuthology.orchestra.run.vm09.stdout:(52/136): boost-program-options-1.75.0-13.el9.x 690 kB/s | 104 kB 00:00 2026-03-10T12:54:59.745 INFO:teuthology.orchestra.run.vm06.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 60 MB/s | 1.6 MB 00:00 2026-03-10T12:54:59.770 INFO:teuthology.orchestra.run.vm09.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 488 kB/s | 15 kB 00:00 2026-03-10T12:54:59.775 INFO:teuthology.orchestra.run.vm04.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 3.3 MB/s | 233 kB 00:00 2026-03-10T12:54:59.780 INFO:teuthology.orchestra.run.vm04.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 4.1 MB/s | 292 kB 00:00 2026-03-10T12:54:59.845 INFO:teuthology.orchestra.run.vm04.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 2.7 MB/s | 188 kB 00:00 2026-03-10T12:54:59.845 INFO:teuthology.orchestra.run.vm04.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 643 kB/s | 42 kB 00:00 2026-03-10T12:54:59.860 INFO:teuthology.orchestra.run.vm09.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.8 MB/s | 164 kB 00:00 2026-03-10T12:54:59.892 INFO:teuthology.orchestra.run.vm09.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 4.9 MB/s | 160 kB 00:00 2026-03-10T12:54:59.925 INFO:teuthology.orchestra.run.vm09.stdout:(56/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.4 MB/s | 45 kB 00:00 2026-03-10T12:54:59.989 INFO:teuthology.orchestra.run.vm09.stdout:(57/136): librdkafka-1.6.1-102.el9.x86_64.rpm 10 MB/s | 662 kB 00:00 2026-03-10T12:55:00.023 INFO:teuthology.orchestra.run.vm09.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 7.3 MB/s | 246 kB 00:00 2026-03-10T12:55:00.057 INFO:teuthology.orchestra.run.vm09.stdout:(59/136): libxslt-1.1.34-12.el9.x86_64.rpm 6.9 MB/s | 233 kB 00:00 2026-03-10T12:55:00.091 INFO:teuthology.orchestra.run.vm09.stdout:(60/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 8.3 MB/s | 292 kB 00:00 2026-03-10T12:55:00.110 INFO:teuthology.orchestra.run.vm04.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 3.8 MB/s | 1.0 MB 00:00 2026-03-10T12:55:00.124 INFO:teuthology.orchestra.run.vm09.stdout:(61/136): lua-5.4.4-4.el9.x86_64.rpm 5.7 MB/s | 188 kB 00:00 2026-03-10T12:55:00.155 INFO:teuthology.orchestra.run.vm09.stdout:(62/136): openblas-0.3.29-1.el9.x86_64.rpm 1.3 MB/s | 42 kB 00:00 2026-03-10T12:55:00.305 INFO:teuthology.orchestra.run.vm09.stdout:(63/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 5.1 MB/s | 3.0 MB 00:00 2026-03-10T12:55:00.444 INFO:teuthology.orchestra.run.vm04.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 8.8 MB/s | 5.3 MB 00:00 2026-03-10T12:55:00.514 INFO:teuthology.orchestra.run.vm09.stdout:(64/136): openblas-openmp-0.3.29-1.el9.x86_64.r 15 MB/s | 5.3 MB 00:00 2026-03-10T12:55:00.517 INFO:teuthology.orchestra.run.vm04.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 3.3 MB/s | 244 kB 00:00 2026-03-10T12:55:00.589 INFO:teuthology.orchestra.run.vm04.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 3.4 MB/s | 249 kB 00:00 2026-03-10T12:55:00.606 INFO:teuthology.orchestra.run.vm04.stdout:(68/136): python3-babel-2.9.1-2.el9.noarch.rpm 12 MB/s | 6.0 MB 00:00 2026-03-10T12:55:00.651 INFO:teuthology.orchestra.run.vm09.stdout:(65/136): protobuf-3.14.0-17.el9.x86_64.rpm 2.9 MB/s | 1.0 MB 00:00 2026-03-10T12:55:00.657 INFO:teuthology.orchestra.run.vm04.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 699 kB/s | 48 kB 00:00 2026-03-10T12:55:00.675 INFO:teuthology.orchestra.run.vm04.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 2.5 MB/s | 177 kB 00:00 2026-03-10T12:55:00.684 INFO:teuthology.orchestra.run.vm09.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 7.2 MB/s | 244 kB 00:00 2026-03-10T12:55:00.718 INFO:teuthology.orchestra.run.vm09.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 7.4 MB/s | 249 kB 00:00 2026-03-10T12:55:00.727 INFO:teuthology.orchestra.run.vm04.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 2.4 MB/s | 172 kB 00:00 2026-03-10T12:55:00.740 INFO:teuthology.orchestra.run.vm04.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 534 kB/s | 35 kB 00:00 2026-03-10T12:55:00.749 INFO:teuthology.orchestra.run.vm09.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 1.5 MB/s | 48 kB 00:00 2026-03-10T12:55:00.782 INFO:teuthology.orchestra.run.vm09.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 5.3 MB/s | 177 kB 00:00 2026-03-10T12:55:00.818 INFO:teuthology.orchestra.run.vm09.stdout:(70/136): python3-mako-1.1.4-6.el9.noarch.rpm 4.7 MB/s | 172 kB 00:00 2026-03-10T12:55:00.851 INFO:teuthology.orchestra.run.vm09.stdout:(71/136): python3-markupsafe-1.1.1-12.el9.x86_6 1.0 MB/s | 35 kB 00:00 2026-03-10T12:55:00.887 INFO:teuthology.orchestra.run.vm09.stdout:(72/136): python3-babel-2.9.1-2.el9.noarch.rpm 16 MB/s | 6.0 MB 00:00 2026-03-10T12:55:00.952 INFO:teuthology.orchestra.run.vm09.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 6.7 MB/s | 442 kB 00:00 2026-03-10T12:55:00.984 INFO:teuthology.orchestra.run.vm09.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 2.4 MB/s | 77 kB 00:00 2026-03-10T12:55:01.018 INFO:teuthology.orchestra.run.vm09.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 7.9 MB/s | 267 kB 00:00 2026-03-10T12:55:01.050 INFO:teuthology.orchestra.run.vm09.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 4.8 MB/s | 157 kB 00:00 2026-03-10T12:55:01.084 INFO:teuthology.orchestra.run.vm09.stdout:(77/136): python3-pyasn1-modules-0.4.8-7.el9.no 8.1 MB/s | 277 kB 00:00 2026-03-10T12:55:01.088 INFO:teuthology.orchestra.run.vm04.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.2 MB/s | 442 kB 00:00 2026-03-10T12:55:01.112 INFO:teuthology.orchestra.run.vm09.stdout:(78/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 24 MB/s | 6.1 MB 00:00 2026-03-10T12:55:01.115 INFO:teuthology.orchestra.run.vm09.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 1.7 MB/s | 54 kB 00:00 2026-03-10T12:55:01.150 INFO:teuthology.orchestra.run.vm09.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 1.2 MB/s | 42 kB 00:00 2026-03-10T12:55:01.154 INFO:teuthology.orchestra.run.vm04.stdout:(74/136): python3-packaging-20.9-5.el9.noarch.r 1.1 MB/s | 77 kB 00:00 2026-03-10T12:55:01.186 INFO:teuthology.orchestra.run.vm09.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 6.6 MB/s | 240 kB 00:00 2026-03-10T12:55:01.218 INFO:teuthology.orchestra.run.vm09.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 1.2 MB/s | 37 kB 00:00 2026-03-10T12:55:01.225 INFO:teuthology.orchestra.run.vm04.stdout:(75/136): python3-protobuf-3.14.0-17.el9.noarch 3.7 MB/s | 267 kB 00:00 2026-03-10T12:55:01.249 INFO:teuthology.orchestra.run.vm09.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.1 MB/s | 66 kB 00:00 2026-03-10T12:55:01.283 INFO:teuthology.orchestra.run.vm09.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 8.8 MB/s | 303 kB 00:00 2026-03-10T12:55:01.298 INFO:teuthology.orchestra.run.vm04.stdout:(76/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 2.1 MB/s | 157 kB 00:00 2026-03-10T12:55:01.315 INFO:teuthology.orchestra.run.vm09.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.0 MB/s | 64 kB 00:00 2026-03-10T12:55:01.391 INFO:teuthology.orchestra.run.vm04.stdout:(77/136): python3-pyasn1-modules-0.4.8-7.el9.no 2.9 MB/s | 277 kB 00:00 2026-03-10T12:55:01.457 INFO:teuthology.orchestra.run.vm04.stdout:(78/136): python3-requests-oauthlib-1.3.0-12.el 812 kB/s | 54 kB 00:00 2026-03-10T12:55:01.508 INFO:teuthology.orchestra.run.vm09.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 116 kB/s | 22 kB 00:00 2026-03-10T12:55:01.572 INFO:teuthology.orchestra.run.vm04.stdout:(79/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 7.3 MB/s | 6.1 MB 00:00 2026-03-10T12:55:01.637 INFO:teuthology.orchestra.run.vm04.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 634 kB/s | 42 kB 00:00 2026-03-10T12:55:01.708 INFO:teuthology.orchestra.run.vm04.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 3.3 MB/s | 240 kB 00:00 2026-03-10T12:55:01.778 INFO:teuthology.orchestra.run.vm04.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 537 kB/s | 37 kB 00:00 2026-03-10T12:55:01.835 INFO:teuthology.orchestra.run.vm09.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 2.6 MB/s | 862 kB 00:00 2026-03-10T12:55:01.846 INFO:teuthology.orchestra.run.vm04.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 970 kB/s | 66 kB 00:00 2026-03-10T12:55:01.853 INFO:teuthology.orchestra.run.vm09.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 31 MB/s | 551 kB 00:00 2026-03-10T12:55:01.861 INFO:teuthology.orchestra.run.vm09.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 41 MB/s | 308 kB 00:00 2026-03-10T12:55:01.863 INFO:teuthology.orchestra.run.vm09.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 10 MB/s | 19 kB 00:00 2026-03-10T12:55:01.934 INFO:teuthology.orchestra.run.vm04.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 3.4 MB/s | 303 kB 00:00 2026-03-10T12:55:01.934 INFO:teuthology.orchestra.run.vm09.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 62 MB/s | 4.4 MB 00:00 2026-03-10T12:55:01.937 INFO:teuthology.orchestra.run.vm09.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-03-10T12:55:01.940 INFO:teuthology.orchestra.run.vm09.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-10T12:55:01.944 INFO:teuthology.orchestra.run.vm09.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 20 MB/s | 67 kB 00:00 2026-03-10T12:55:01.948 INFO:teuthology.orchestra.run.vm09.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 38 MB/s | 151 kB 00:00 2026-03-10T12:55:01.961 INFO:teuthology.orchestra.run.vm09.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 68 MB/s | 838 kB 00:00 2026-03-10T12:55:01.974 INFO:teuthology.orchestra.run.vm09.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 43 MB/s | 548 kB 00:00 2026-03-10T12:55:01.976 INFO:teuthology.orchestra.run.vm09.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 12 MB/s | 29 kB 00:00 2026-03-10T12:55:01.980 INFO:teuthology.orchestra.run.vm09.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 18 MB/s | 60 kB 00:00 2026-03-10T12:55:01.982 INFO:teuthology.orchestra.run.vm09.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 19 MB/s | 43 kB 00:00 2026-03-10T12:55:01.985 INFO:teuthology.orchestra.run.vm09.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 14 MB/s | 32 kB 00:00 2026-03-10T12:55:01.987 INFO:teuthology.orchestra.run.vm09.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.9 MB/s | 14 kB 00:00 2026-03-10T12:55:01.991 INFO:teuthology.orchestra.run.vm09.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 40 MB/s | 173 kB 00:00 2026-03-10T12:55:01.998 INFO:teuthology.orchestra.run.vm09.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 57 MB/s | 358 kB 00:00 2026-03-10T12:55:02.000 INFO:teuthology.orchestra.run.vm04.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 966 kB/s | 64 kB 00:00 2026-03-10T12:55:02.003 INFO:teuthology.orchestra.run.vm09.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 52 MB/s | 254 kB 00:00 2026-03-10T12:55:02.054 INFO:teuthology.orchestra.run.vm09.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 41 MB/s | 2.0 MB 00:00 2026-03-10T12:55:02.059 INFO:teuthology.orchestra.run.vm09.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 28 MB/s | 144 kB 00:00 2026-03-10T12:55:02.062 INFO:teuthology.orchestra.run.vm09.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 4.9 MB/s | 11 kB 00:00 2026-03-10T12:55:02.064 INFO:teuthology.orchestra.run.vm09.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 8.1 MB/s | 18 kB 00:00 2026-03-10T12:55:02.066 INFO:teuthology.orchestra.run.vm09.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 10 MB/s | 23 kB 00:00 2026-03-10T12:55:02.069 INFO:teuthology.orchestra.run.vm09.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.4 MB/s | 20 kB 00:00 2026-03-10T12:55:02.071 INFO:teuthology.orchestra.run.vm09.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 9.8 MB/s | 19 kB 00:00 2026-03-10T12:55:02.073 INFO:teuthology.orchestra.run.vm09.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-10T12:55:02.097 INFO:teuthology.orchestra.run.vm09.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 43 MB/s | 1.0 MB 00:00 2026-03-10T12:55:02.124 INFO:teuthology.orchestra.run.vm09.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 1.7 MB/s | 46 kB 00:00 2026-03-10T12:55:02.124 INFO:teuthology.orchestra.run.vm06.stdout:(134/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 3.2 MB 00:02 2026-03-10T12:55:02.126 INFO:teuthology.orchestra.run.vm09.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 29 MB/s | 79 kB 00:00 2026-03-10T12:55:02.130 INFO:teuthology.orchestra.run.vm09.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 16 MB/s | 58 kB 00:00 2026-03-10T12:55:02.136 INFO:teuthology.orchestra.run.vm09.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 48 MB/s | 272 kB 00:00 2026-03-10T12:55:02.140 INFO:teuthology.orchestra.run.vm09.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.4 MB/s | 16 kB 00:00 2026-03-10T12:55:02.144 INFO:teuthology.orchestra.run.vm09.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 21 MB/s | 90 kB 00:00 2026-03-10T12:55:02.146 INFO:teuthology.orchestra.run.vm09.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 15 MB/s | 31 kB 00:00 2026-03-10T12:55:02.150 INFO:teuthology.orchestra.run.vm09.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 48 MB/s | 188 kB 00:00 2026-03-10T12:55:02.153 INFO:teuthology.orchestra.run.vm09.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 23 MB/s | 59 kB 00:00 2026-03-10T12:55:02.156 INFO:teuthology.orchestra.run.vm09.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 11 MB/s | 36 kB 00:00 2026-03-10T12:55:02.159 INFO:teuthology.orchestra.run.vm09.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 30 MB/s | 86 kB 00:00 2026-03-10T12:55:02.165 INFO:teuthology.orchestra.run.vm09.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 42 MB/s | 230 kB 00:00 2026-03-10T12:55:02.165 INFO:teuthology.orchestra.run.vm04.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 135 kB/s | 22 kB 00:00 2026-03-10T12:55:02.168 INFO:teuthology.orchestra.run.vm09.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 25 MB/s | 90 kB 00:00 2026-03-10T12:55:02.182 INFO:teuthology.orchestra.run.vm09.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 32 MB/s | 427 kB 00:00 2026-03-10T12:55:02.186 INFO:teuthology.orchestra.run.vm09.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 4.9 MB/s | 22 kB 00:00 2026-03-10T12:55:02.194 INFO:teuthology.orchestra.run.vm09.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 2.7 MB/s | 20 kB 00:00 2026-03-10T12:55:02.198 INFO:teuthology.orchestra.run.vm09.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 44 MB/s | 191 kB 00:00 2026-03-10T12:55:02.237 INFO:teuthology.orchestra.run.vm09.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 41 MB/s | 1.6 MB 00:00 2026-03-10T12:55:02.324 INFO:teuthology.orchestra.run.vm04.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 5.3 MB/s | 862 kB 00:00 2026-03-10T12:55:02.339 INFO:teuthology.orchestra.run.vm04.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 38 MB/s | 551 kB 00:00 2026-03-10T12:55:02.345 INFO:teuthology.orchestra.run.vm04.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 49 MB/s | 308 kB 00:00 2026-03-10T12:55:02.347 INFO:teuthology.orchestra.run.vm04.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 9.5 MB/s | 19 kB 00:00 2026-03-10T12:55:02.415 INFO:teuthology.orchestra.run.vm04.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 65 MB/s | 4.4 MB 00:00 2026-03-10T12:55:02.440 INFO:teuthology.orchestra.run.vm04.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 1.0 MB/s | 25 kB 00:00 2026-03-10T12:55:02.440 INFO:teuthology.orchestra.run.vm09.stdout:(133/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 15 MB/s | 19 MB 00:01 2026-03-10T12:55:02.442 INFO:teuthology.orchestra.run.vm04.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 20 MB/s | 49 kB 00:00 2026-03-10T12:55:02.445 INFO:teuthology.orchestra.run.vm04.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 23 MB/s | 67 kB 00:00 2026-03-10T12:55:02.450 INFO:teuthology.orchestra.run.vm04.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 36 MB/s | 151 kB 00:00 2026-03-10T12:55:02.461 INFO:teuthology.orchestra.run.vm04.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 72 MB/s | 838 kB 00:00 2026-03-10T12:55:02.470 INFO:teuthology.orchestra.run.vm04.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 60 MB/s | 548 kB 00:00 2026-03-10T12:55:02.472 INFO:teuthology.orchestra.run.vm04.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 14 MB/s | 29 kB 00:00 2026-03-10T12:55:02.556 INFO:teuthology.orchestra.run.vm04.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 723 kB/s | 60 kB 00:00 2026-03-10T12:55:02.657 INFO:teuthology.orchestra.run.vm04.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 432 kB/s | 43 kB 00:00 2026-03-10T12:55:02.677 INFO:teuthology.orchestra.run.vm04.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 1.5 MB/s | 32 kB 00:00 2026-03-10T12:55:02.683 INFO:teuthology.orchestra.run.vm04.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 2.4 MB/s | 14 kB 00:00 2026-03-10T12:55:02.688 INFO:teuthology.orchestra.run.vm04.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 39 MB/s | 173 kB 00:00 2026-03-10T12:55:02.694 INFO:teuthology.orchestra.run.vm04.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 62 MB/s | 358 kB 00:00 2026-03-10T12:55:02.719 INFO:teuthology.orchestra.run.vm04.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 9.9 MB/s | 254 kB 00:00 2026-03-10T12:55:02.792 INFO:teuthology.orchestra.run.vm04.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 28 MB/s | 2.0 MB 00:00 2026-03-10T12:55:02.816 INFO:teuthology.orchestra.run.vm04.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 6.0 MB/s | 144 kB 00:00 2026-03-10T12:55:02.819 INFO:teuthology.orchestra.run.vm04.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 3.3 MB/s | 11 kB 00:00 2026-03-10T12:55:02.821 INFO:teuthology.orchestra.run.vm04.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 8.8 MB/s | 18 kB 00:00 2026-03-10T12:55:02.824 INFO:teuthology.orchestra.run.vm04.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 10 MB/s | 23 kB 00:00 2026-03-10T12:55:02.826 INFO:teuthology.orchestra.run.vm04.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.8 MB/s | 20 kB 00:00 2026-03-10T12:55:02.828 INFO:teuthology.orchestra.run.vm04.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 9.1 MB/s | 19 kB 00:00 2026-03-10T12:55:02.830 INFO:teuthology.orchestra.run.vm04.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 13 MB/s | 26 kB 00:00 2026-03-10T12:55:02.845 INFO:teuthology.orchestra.run.vm04.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 73 MB/s | 1.0 MB 00:00 2026-03-10T12:55:02.848 INFO:teuthology.orchestra.run.vm04.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 14 MB/s | 46 kB 00:00 2026-03-10T12:55:02.850 INFO:teuthology.orchestra.run.vm04.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 30 MB/s | 79 kB 00:00 2026-03-10T12:55:02.854 INFO:teuthology.orchestra.run.vm04.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 19 MB/s | 58 kB 00:00 2026-03-10T12:55:02.860 INFO:teuthology.orchestra.run.vm04.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 45 MB/s | 272 kB 00:00 2026-03-10T12:55:02.864 INFO:teuthology.orchestra.run.vm04.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.2 MB/s | 16 kB 00:00 2026-03-10T12:55:02.867 INFO:teuthology.orchestra.run.vm04.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 25 MB/s | 90 kB 00:00 2026-03-10T12:55:02.870 INFO:teuthology.orchestra.run.vm04.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 14 MB/s | 31 kB 00:00 2026-03-10T12:55:02.874 INFO:teuthology.orchestra.run.vm04.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 49 MB/s | 188 kB 00:00 2026-03-10T12:55:02.876 INFO:teuthology.orchestra.run.vm04.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 25 MB/s | 59 kB 00:00 2026-03-10T12:55:02.878 INFO:teuthology.orchestra.run.vm04.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 17 MB/s | 36 kB 00:00 2026-03-10T12:55:02.881 INFO:teuthology.orchestra.run.vm04.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 32 MB/s | 86 kB 00:00 2026-03-10T12:55:02.886 INFO:teuthology.orchestra.run.vm04.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 45 MB/s | 230 kB 00:00 2026-03-10T12:55:02.890 INFO:teuthology.orchestra.run.vm04.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 28 MB/s | 90 kB 00:00 2026-03-10T12:55:02.899 INFO:teuthology.orchestra.run.vm04.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 48 MB/s | 427 kB 00:00 2026-03-10T12:55:02.902 INFO:teuthology.orchestra.run.vm04.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 9.1 MB/s | 22 kB 00:00 2026-03-10T12:55:02.904 INFO:teuthology.orchestra.run.vm04.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 7.8 MB/s | 20 kB 00:00 2026-03-10T12:55:02.909 INFO:teuthology.orchestra.run.vm04.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 49 MB/s | 191 kB 00:00 2026-03-10T12:55:02.932 INFO:teuthology.orchestra.run.vm04.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 69 MB/s | 1.6 MB 00:00 2026-03-10T12:55:03.913 INFO:teuthology.orchestra.run.vm09.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.0 MB/s | 3.4 MB 00:01 2026-03-10T12:55:03.913 INFO:teuthology.orchestra.run.vm06.stdout:(135/136): librados2-19.2.3-678.ge911bdeb.el9.x 839 kB/s | 3.4 MB 00:04 2026-03-10T12:55:04.649 INFO:teuthology.orchestra.run.vm09.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.4 MB/s | 3.2 MB 00:02 2026-03-10T12:55:04.732 INFO:teuthology.orchestra.run.vm04.stdout:(133/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 5.9 MB/s | 19 MB 00:03 2026-03-10T12:55:05.059 INFO:teuthology.orchestra.run.vm04.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 1.6 MB/s | 3.4 MB 00:02 2026-03-10T12:55:07.087 INFO:teuthology.orchestra.run.vm04.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.3 MB/s | 3.2 MB 00:02 2026-03-10T12:55:17.156 INFO:teuthology.orchestra.run.vm06.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 1.7 MB/s | 50 MB 00:28 2026-03-10T12:55:17.161 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:55:17.161 INFO:teuthology.orchestra.run.vm06.stdout:Total 5.0 MB/s | 210 MB 00:42 2026-03-10T12:55:17.725 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T12:55:17.776 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T12:55:17.776 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T12:55:18.617 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T12:55:18.617 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T12:55:19.156 INFO:teuthology.orchestra.run.vm04.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 1.7 MB/s | 50 MB 00:29 2026-03-10T12:55:19.158 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:55:19.159 INFO:teuthology.orchestra.run.vm04.stdout:Total 4.9 MB/s | 210 MB 00:43 2026-03-10T12:55:19.547 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T12:55:19.562 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-10T12:55:19.577 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-10T12:55:19.752 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T12:55:19.765 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-10T12:55:19.767 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:19.801 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T12:55:19.801 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T12:55:19.828 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:19.829 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:19.860 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:19.869 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:19.873 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-10T12:55:19.876 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-10T12:55:19.881 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-10T12:55:19.891 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-10T12:55:19.894 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:19.934 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:19.935 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:19.949 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:19.984 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-10T12:55:20.066 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-10T12:55:20.232 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-10T12:55:20.290 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-10T12:55:20.316 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-10T12:55:20.327 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-10T12:55:20.339 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-10T12:55:20.348 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-10T12:55:20.355 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-10T12:55:20.360 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-10T12:55:20.388 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-10T12:55:20.407 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-10T12:55:20.413 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-10T12:55:20.420 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-10T12:55:20.423 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-10T12:55:20.454 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-10T12:55:20.461 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-10T12:55:20.473 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-10T12:55:20.488 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-10T12:55:20.522 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-10T12:55:20.568 INFO:teuthology.orchestra.run.vm09.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 1.7 MB/s | 50 MB 00:28 2026-03-10T12:55:20.571 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:55:20.571 INFO:teuthology.orchestra.run.vm09.stdout:Total 4.6 MB/s | 210 MB 00:45 2026-03-10T12:55:20.599 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-10T12:55:20.606 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-10T12:55:20.616 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-10T12:55:20.628 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T12:55:20.628 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T12:55:20.650 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-10T12:55:20.712 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-10T12:55:20.730 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-10T12:55:20.738 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-10T12:55:20.748 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-10T12:55:20.755 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-10T12:55:20.760 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-10T12:55:20.778 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-10T12:55:20.805 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-10T12:55:20.813 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-10T12:55:20.822 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-10T12:55:20.836 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-10T12:55:20.855 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-10T12:55:20.868 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-10T12:55:20.940 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-10T12:55:20.950 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-10T12:55:20.960 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-10T12:55:21.010 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-10T12:55:21.203 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T12:55:21.254 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T12:55:21.254 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T12:55:21.392 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-10T12:55:21.411 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-10T12:55:21.419 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-10T12:55:21.427 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-10T12:55:21.433 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-10T12:55:21.441 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-10T12:55:21.446 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-10T12:55:21.449 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-10T12:55:21.481 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-10T12:55:21.539 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-10T12:55:21.551 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T12:55:21.556 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-10T12:55:21.565 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-10T12:55:21.567 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-10T12:55:21.573 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-10T12:55:21.580 INFO:teuthology.orchestra.run.vm04.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-10T12:55:21.583 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-10T12:55:21.590 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-10T12:55:21.600 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-10T12:55:21.606 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-10T12:55:21.640 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-10T12:55:21.656 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-10T12:55:21.700 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-10T12:55:21.759 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-10T12:55:21.762 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:21.825 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:21.827 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:21.858 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:21.870 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:21.874 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-10T12:55:21.877 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-10T12:55:21.884 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-10T12:55:21.894 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-10T12:55:21.896 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:21.934 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:21.936 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:21.953 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:21.973 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-10T12:55:21.990 INFO:teuthology.orchestra.run.vm04.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-10T12:55:22.006 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-10T12:55:22.013 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-10T12:55:22.030 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-10T12:55:22.037 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-10T12:55:22.062 INFO:teuthology.orchestra.run.vm04.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-10T12:55:22.077 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-10T12:55:22.078 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-10T12:55:22.083 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-10T12:55:22.087 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-10T12:55:22.099 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-10T12:55:22.102 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T12:55:22.102 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T12:55:22.107 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-10T12:55:22.108 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-10T12:55:22.112 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-10T12:55:22.118 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-10T12:55:22.147 INFO:teuthology.orchestra.run.vm04.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-10T12:55:22.165 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-10T12:55:22.170 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-10T12:55:22.178 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-10T12:55:22.181 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-10T12:55:22.213 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-10T12:55:22.221 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-10T12:55:22.233 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-10T12:55:22.248 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-10T12:55:22.257 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-10T12:55:22.289 INFO:teuthology.orchestra.run.vm04.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-10T12:55:22.296 INFO:teuthology.orchestra.run.vm04.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-10T12:55:22.305 INFO:teuthology.orchestra.run.vm04.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-10T12:55:22.339 INFO:teuthology.orchestra.run.vm04.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-10T12:55:22.400 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-10T12:55:22.418 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-10T12:55:22.427 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-10T12:55:22.437 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-10T12:55:22.444 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-10T12:55:22.449 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-10T12:55:22.466 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-10T12:55:22.493 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-10T12:55:22.501 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-10T12:55:22.506 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-10T12:55:22.508 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-10T12:55:22.523 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-10T12:55:22.536 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-10T12:55:22.548 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-10T12:55:22.607 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-10T12:55:22.613 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-10T12:55:22.635 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-10T12:55:22.647 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-10T12:55:22.695 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-10T12:55:23.040 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T12:55:23.054 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-10T12:55:23.066 INFO:teuthology.orchestra.run.vm09.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-10T12:55:23.074 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-10T12:55:23.094 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-10T12:55:23.100 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-10T12:55:23.109 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-10T12:55:23.114 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-10T12:55:23.122 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-10T12:55:23.127 INFO:teuthology.orchestra.run.vm04.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-10T12:55:23.130 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-10T12:55:23.161 INFO:teuthology.orchestra.run.vm04.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-10T12:55:23.215 INFO:teuthology.orchestra.run.vm04.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-10T12:55:23.230 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-10T12:55:23.237 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-10T12:55:23.239 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:23.240 INFO:teuthology.orchestra.run.vm04.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-10T12:55:23.247 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-10T12:55:23.256 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-10T12:55:23.318 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-10T12:55:23.329 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-10T12:55:23.335 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-10T12:55:23.362 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:23.365 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:23.371 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-10T12:55:23.390 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-10T12:55:23.396 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-10T12:55:23.407 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:23.410 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-10T12:55:23.413 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-10T12:55:23.419 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-10T12:55:23.427 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-10T12:55:23.428 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-10T12:55:23.430 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:23.434 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-10T12:55:23.460 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-10T12:55:23.466 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-10T12:55:23.471 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-10T12:55:23.471 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:23.473 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:23.490 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-10T12:55:23.526 INFO:teuthology.orchestra.run.vm09.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-10T12:55:23.562 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-10T12:55:23.568 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-10T12:55:23.592 INFO:teuthology.orchestra.run.vm09.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-10T12:55:23.607 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-10T12:55:23.616 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-10T12:55:23.626 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-10T12:55:23.631 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-10T12:55:23.633 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-10T12:55:23.634 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:23.637 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-10T12:55:23.698 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-10T12:55:23.705 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-10T12:55:23.746 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:23.750 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-10T12:55:23.754 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-10T12:55:23.754 INFO:teuthology.orchestra.run.vm09.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-10T12:55:23.758 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-10T12:55:23.762 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-10T12:55:23.773 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-10T12:55:23.777 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-10T12:55:23.785 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-10T12:55:23.787 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-10T12:55:23.817 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-10T12:55:23.824 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-10T12:55:23.829 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-10T12:55:23.832 INFO:teuthology.orchestra.run.vm04.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-10T12:55:23.835 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-10T12:55:23.848 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-10T12:55:23.856 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-10T12:55:23.857 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-10T12:55:23.887 INFO:teuthology.orchestra.run.vm09.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-10T12:55:23.893 INFO:teuthology.orchestra.run.vm09.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-10T12:55:23.901 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-10T12:55:23.930 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-10T12:55:23.990 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-10T12:55:24.006 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-10T12:55:24.014 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-10T12:55:24.022 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-10T12:55:24.024 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-10T12:55:24.025 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:24.031 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-10T12:55:24.035 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-10T12:55:24.049 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:24.052 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-10T12:55:24.057 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-10T12:55:24.083 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-10T12:55:24.090 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-10T12:55:24.097 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-10T12:55:24.111 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-10T12:55:24.123 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-10T12:55:24.135 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-10T12:55:24.199 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-10T12:55:24.216 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-10T12:55:24.226 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-10T12:55:24.239 INFO:teuthology.orchestra.run.vm04.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-10T12:55:24.273 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-10T12:55:24.330 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-10T12:55:24.652 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-10T12:55:24.669 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-10T12:55:24.675 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-10T12:55:24.683 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-10T12:55:24.688 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-10T12:55:24.697 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-10T12:55:24.702 INFO:teuthology.orchestra.run.vm09.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-10T12:55:24.704 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-10T12:55:24.734 INFO:teuthology.orchestra.run.vm09.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-10T12:55:24.786 INFO:teuthology.orchestra.run.vm09.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-10T12:55:24.801 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-10T12:55:24.810 INFO:teuthology.orchestra.run.vm09.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-10T12:55:24.815 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-10T12:55:24.824 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-10T12:55:24.830 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-10T12:55:24.839 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-10T12:55:24.845 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-10T12:55:24.882 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-10T12:55:24.896 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-10T12:55:24.940 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-10T12:55:25.132 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-10T12:55:25.165 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-10T12:55:25.173 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-10T12:55:25.179 INFO:teuthology.orchestra.run.vm04.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-10T12:55:25.191 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:25.197 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:25.212 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-10T12:55:25.220 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:25.238 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-10T12:55:25.244 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-10T12:55:25.252 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-10T12:55:25.258 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-10T12:55:25.316 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-10T12:55:25.319 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-10T12:55:25.341 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-10T12:55:25.344 INFO:teuthology.orchestra.run.vm04.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:25.345 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-10T12:55:25.352 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-10T12:55:25.366 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-10T12:55:25.380 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:25.386 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-10T12:55:25.394 INFO:teuthology.orchestra.run.vm04.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-10T12:55:25.395 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-10T12:55:25.440 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-10T12:55:25.502 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-10T12:55:25.513 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-10T12:55:25.520 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:25.526 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-10T12:55:25.531 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-10T12:55:25.534 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:25.552 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:25.659 INFO:teuthology.orchestra.run.vm04.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-10T12:55:25.663 INFO:teuthology.orchestra.run.vm04.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:25.682 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:25.685 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-10T12:55:25.743 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-10T12:55:25.833 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-10T12:55:25.863 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-10T12:55:25.869 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:25.910 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:25.911 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-10T12:55:25.911 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T12:55:25.911 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:25.916 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:26.610 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-10T12:55:26.641 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-10T12:55:26.647 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-10T12:55:26.654 INFO:teuthology.orchestra.run.vm09.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-10T12:55:26.812 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-10T12:55:26.815 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:26.824 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:26.832 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:26.851 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-10T12:55:26.854 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:26.856 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-10T12:55:26.864 INFO:teuthology.orchestra.run.vm09.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-10T12:55:26.871 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-10T12:55:26.891 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-10T12:55:26.980 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-10T12:55:26.995 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-10T12:55:27.023 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-10T12:55:27.061 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-10T12:55:27.113 INFO:teuthology.orchestra.run.vm09.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-10T12:55:27.116 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:27.123 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-10T12:55:27.133 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-10T12:55:27.137 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-10T12:55:27.140 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-10T12:55:27.140 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:27.146 INFO:teuthology.orchestra.run.vm04.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-10T12:55:27.151 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-10T12:55:27.180 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:27.198 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:27.515 INFO:teuthology.orchestra.run.vm04.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-10T12:55:27.521 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:27.562 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:27.562 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-10T12:55:27.562 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T12:55:27.562 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:27.568 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:28.294 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:28.299 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:28.323 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-10T12:55:28.340 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-10T12:55:28.360 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-10T12:55:28.450 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-10T12:55:28.465 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-10T12:55:28.493 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-10T12:55:28.531 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-10T12:55:28.597 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-10T12:55:28.607 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-10T12:55:28.613 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:28.620 INFO:teuthology.orchestra.run.vm09.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-10T12:55:28.625 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-10T12:55:28.628 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:28.647 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-10T12:55:28.961 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-10T12:55:28.966 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:29.012 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-10T12:55:29.012 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-10T12:55:29.012 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-10T12:55:29.012 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:29.018 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-10T12:55:32.486 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:32.762 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:32.785 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:33.266 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:33.287 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:33.356 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-10T12:55:33.362 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-10T12:55:33.390 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:33.390 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'qat' with GID 994. 2026-03-10T12:55:33.390 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T12:55:33.390 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T12:55:33.390 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:33.403 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:33.430 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:33.430 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T12:55:33.431 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:33.471 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-10T12:55:33.551 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-10T12:55:33.556 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:33.570 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:33.570 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:33.570 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T12:55:33.570 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:33.848 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:33.848 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-10T12:55:33.848 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-10T12:55:33.848 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-10T12:55:33.848 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-10T12:55:33.849 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-10T12:55:33.849 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-10T12:55:33.849 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-10T12:55:33.849 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:33.975 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:34.000 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:34.001 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:34.001 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T12:55:34.001 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:34.001 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:34.001 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:34.308 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:34.328 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:34.434 INFO:teuthology.orchestra.run.vm04.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-10T12:55:34.489 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:34.513 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:34.533 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-10T12:55:34.829 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:34.829 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'qat' with GID 994. 2026-03-10T12:55:34.829 INFO:teuthology.orchestra.run.vm04.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T12:55:34.829 INFO:teuthology.orchestra.run.vm04.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T12:55:34.829 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:34.947 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:34.979 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:34.979 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T12:55:34.979 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:34.981 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:34.988 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:34.995 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-10T12:55:35.019 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-10T12:55:35.022 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:35.023 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-10T12:55:35.100 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-10T12:55:35.106 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:35.119 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:35.119 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:35.119 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T12:55:35.119 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /sys 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /proc 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /mnt 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /var/tmp 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /home 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /root 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /tmp 2026-03-10T12:55:35.447 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:35.566 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:35.574 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:35.575 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-10T12:55:35.600 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:35.833 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-10T12:55:35.857 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:35.865 INFO:teuthology.orchestra.run.vm09.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-10T12:55:35.868 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-10T12:55:35.887 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:35.887 INFO:teuthology.orchestra.run.vm09.stdout:Creating group 'qat' with GID 994. 2026-03-10T12:55:35.887 INFO:teuthology.orchestra.run.vm09.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-10T12:55:35.887 INFO:teuthology.orchestra.run.vm09.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-10T12:55:35.887 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:35.897 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:35.922 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-10T12:55:35.922 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-10T12:55:35.922 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:35.945 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:35.970 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-10T12:55:36.032 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:36.036 INFO:teuthology.orchestra.run.vm04.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:36.042 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-10T12:55:36.047 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-10T12:55:36.052 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:36.065 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-10T12:55:36.067 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-10T12:55:36.067 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:36.067 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T12:55:36.067 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:36.069 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:36.107 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:36.109 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:36.173 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:36.232 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-10T12:55:36.235 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:36.260 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:36.274 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:36.286 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:36.630 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:36.638 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:36.821 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-10T12:55:36.825 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:36.850 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:36.862 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:36.876 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:36.885 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:36.885 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:36.885 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T12:55:36.885 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:36.903 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-10T12:55:36.903 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:36.904 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T12:55:36.904 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:36.904 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-10T12:55:36.904 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:36.967 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:36.971 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-10T12:55:36.977 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-10T12:55:37.001 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-10T12:55:37.005 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:37.050 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:37.072 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:37.174 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:37.177 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:37.246 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:37.309 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-10T12:55:37.311 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:37.336 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:37.353 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:37.368 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:37.556 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-10T12:55:37.564 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:37.895 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-10T12:55:37.899 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:37.921 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:37.933 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:37.953 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:37.953 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:37.953 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T12:55:37.953 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:38.092 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-10T12:55:38.094 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:38.111 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:38.132 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:38.155 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-10T12:55:38.212 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-10T12:55:38.215 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-10T12:55:38.238 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:38.252 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:38.263 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-10T12:55:38.771 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-10T12:55:38.775 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-10T12:55:38.797 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:38.809 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:38.830 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-10T12:55:38.830 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:38.830 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T12:55:38.830 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:38.992 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-10T12:55:39.013 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:39.712 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-10T12:55:39.723 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-10T12:55:39.729 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-10T12:55:39.786 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-10T12:55:39.796 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:39.800 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-10T12:55:39.800 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:39.815 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:39.815 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:40.946 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-10T12:55:41.000 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-10T12:55:41.006 INFO:teuthology.orchestra.run.vm04.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-10T12:55:41.210 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-10T12:55:41.211 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-10T12:55:41.212 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-10T12:55:41.214 INFO:teuthology.orchestra.run.vm04.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-10T12:55:41.225 INFO:teuthology.orchestra.run.vm04.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:41.230 INFO:teuthology.orchestra.run.vm04.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-10T12:55:41.230 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:41.247 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:41.247 INFO:teuthology.orchestra.run.vm04.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:41.318 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.319 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T12:55:41.320 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:55:41.321 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T12:55:41.413 DEBUG:teuthology.parallel:result is None 2026-03-10T12:55:41.637 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-10T12:55:41.648 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-10T12:55:41.654 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-10T12:55:41.713 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-10T12:55:41.814 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:41.865 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-10T12:55:41.865 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:41.884 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-10T12:55:41.884 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-10T12:55:42.780 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-10T12:55:42.781 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-10T12:55:42.783 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-10T12:55:42.784 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-10T12:55:42.785 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-10T12:55:43.203 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout:Upgraded: 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout:Installed: 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T12:55:43.204 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T12:55:43.205 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:55:43.206 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T12:55:43.491 DEBUG:teuthology.parallel:result is None 2026-03-10T12:55:43.574 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-10T12:55:43.575 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-10T12:55:43.576 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-10T12:55:43.577 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-10T12:55:43.578 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-10T12:55:43.579 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-10T12:55:43.580 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout:Upgraded: 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.694 INFO:teuthology.orchestra.run.vm09.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T12:55:43.695 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-03-10T12:55:43.696 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:55:43.697 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T12:55:43.792 DEBUG:teuthology.parallel:result is None 2026-03-10T12:55:43.793 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:44.445 DEBUG:teuthology.orchestra.run.vm04:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T12:55:44.464 INFO:teuthology.orchestra.run.vm04.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:44.464 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:44.464 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T12:55:44.465 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:45.054 DEBUG:teuthology.orchestra.run.vm06:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T12:55:45.077 INFO:teuthology.orchestra.run.vm06.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:45.077 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:45.077 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T12:55:45.078 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:45.712 DEBUG:teuthology.orchestra.run.vm09:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-10T12:55:45.730 INFO:teuthology.orchestra.run.vm09.stdout:19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:45.730 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-10T12:55:45.730 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-10T12:55:45.731 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-10T12:55:45.731 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:55:45.731 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T12:55:45.758 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:55:45.758 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T12:55:45.783 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:55:45.783 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-10T12:55:45.809 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-10T12:55:45.809 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:55:45.809 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T12:55:45.832 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T12:55:45.895 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:55:45.895 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T12:55:45.918 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T12:55:45.981 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:55:45.981 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/daemon-helper 2026-03-10T12:55:46.004 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-10T12:55:46.067 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-10T12:55:46.067 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:55:46.067 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T12:55:46.090 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T12:55:46.154 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:55:46.154 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T12:55:46.180 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T12:55:46.242 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:55:46.242 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-10T12:55:46.269 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-10T12:55:46.334 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-10T12:55:46.334 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:55:46.334 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T12:55:46.359 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T12:55:46.421 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:55:46.421 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T12:55:46.445 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T12:55:46.509 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:55:46.509 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/stdin-killer 2026-03-10T12:55:46.534 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-10T12:55:46.599 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'mons down', 'mon down', 'out of quorum', 'CEPHADM_STRAY_HOST', 'CEPHADM_STRAY_DAEMON', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Cluster fsid is 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.104', 'mon.b': '192.168.123.106', 'mon.c': '192.168.123.109'} 2026-03-10T12:55:46.643 INFO:tasks.cephadm:First mon is mon.a on vm04 2026-03-10T12:55:46.643 INFO:tasks.cephadm:First mgr is a 2026-03-10T12:55:46.643 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T12:55:46.643 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-10T12:55:46.667 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-10T12:55:46.691 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-10T12:55:46.717 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T12:55:46.717 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:47.293 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T12:55:47.947 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:55:47.948 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T12:55:47.948 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T12:55:47.948 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:49.468 INFO:teuthology.orchestra.run.vm04.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 12:55 /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:49.468 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:50.953 INFO:teuthology.orchestra.run.vm06.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 12:55 /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:50.953 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:52.414 INFO:teuthology.orchestra.run.vm09.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 12:55 /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:52.414 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:52.431 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:52.449 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T12:55:52.471 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T12:55:52.471 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T12:55:52.473 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T12:55:52.492 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T12:55:52.648 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:55:52.664 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:55:52.685 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:56:28.852 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-10T12:56:28.853 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-10T12:56:31.215 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-10T12:57:08.959 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-10T12:57:08.980 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-10T12:57:09.006 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-10T12:57:09.034 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-10T12:57:09.065 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-10T12:57:09.092 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-10T12:57:09.118 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-10T12:57:09.144 INFO:tasks.cephadm:Writing seed config... 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T12:57:09.145 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T12:57:09.145 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:57:09.145 DEBUG:teuthology.orchestra.run.vm04:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T12:57:09.162 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 74d0e8a6-1c80-11f1-9542-4d7926099318 mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T12:57:09.162 DEBUG:teuthology.orchestra.run.vm04:mon.a> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service 2026-03-10T12:57:09.204 DEBUG:teuthology.orchestra.run.vm04:mgr.a> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a.service 2026-03-10T12:57:09.245 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T12:57:09.245 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.104 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:09.386 INFO:teuthology.orchestra.run.vm04.stdout:-------------------------------------------------------------------------------- 2026-03-10T12:57:09.386 INFO:teuthology.orchestra.run.vm04.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '74d0e8a6-1c80-11f1-9542-4d7926099318', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.104', '--skip-admin-label'] 2026-03-10T12:57:09.386 INFO:teuthology.orchestra.run.vm04.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T12:57:09.386 INFO:teuthology.orchestra.run.vm04.stdout:Verifying podman|docker is present... 2026-03-10T12:57:09.405 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stdout 5.8.0 2026-03-10T12:57:09.405 INFO:teuthology.orchestra.run.vm04.stdout:Verifying lvm2 is present... 2026-03-10T12:57:09.405 INFO:teuthology.orchestra.run.vm04.stdout:Verifying time synchronization is in place... 2026-03-10T12:57:09.413 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T12:57:09.413 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T12:57:09.419 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T12:57:09.419 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout inactive 2026-03-10T12:57:09.426 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout enabled 2026-03-10T12:57:09.432 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout active 2026-03-10T12:57:09.432 INFO:teuthology.orchestra.run.vm04.stdout:Unit chronyd.service is enabled and running 2026-03-10T12:57:09.432 INFO:teuthology.orchestra.run.vm04.stdout:Repeating the final host check... 2026-03-10T12:57:09.452 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stdout 5.8.0 2026-03-10T12:57:09.452 INFO:teuthology.orchestra.run.vm04.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-10T12:57:09.452 INFO:teuthology.orchestra.run.vm04.stdout:systemctl is present 2026-03-10T12:57:09.452 INFO:teuthology.orchestra.run.vm04.stdout:lvcreate is present 2026-03-10T12:57:09.459 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T12:57:09.460 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T12:57:09.466 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T12:57:09.466 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout inactive 2026-03-10T12:57:09.473 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout enabled 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stdout active 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Unit chronyd.service is enabled and running 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Host looks OK 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Cluster fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Acquiring lock 140365651759888 on /run/cephadm/74d0e8a6-1c80-11f1-9542-4d7926099318.lock 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Lock 140365651759888 acquired on /run/cephadm/74d0e8a6-1c80-11f1-9542-4d7926099318.lock 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Verifying IP 192.168.123.104 port 3300 ... 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Verifying IP 192.168.123.104 port 6789 ... 2026-03-10T12:57:09.478 INFO:teuthology.orchestra.run.vm04.stdout:Base mon IP(s) is [192.168.123.104:3300, 192.168.123.104:6789], mon addrv is [v2:192.168.123.104:3300,v1:192.168.123.104:6789] 2026-03-10T12:57:09.481 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.104 metric 100 2026-03-10T12:57:09.481 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.104 metric 100 2026-03-10T12:57:09.483 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T12:57:09.483 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:4/64 scope link noprefixroute 2026-03-10T12:57:09.485 INFO:teuthology.orchestra.run.vm04.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T12:57:09.486 INFO:teuthology.orchestra.run.vm04.stdout:Mon IP `192.168.123.104` is in CIDR network `192.168.123.0/24` 2026-03-10T12:57:09.486 INFO:teuthology.orchestra.run.vm04.stdout:Mon IP `192.168.123.104` is in CIDR network `192.168.123.0/24` 2026-03-10T12:57:09.486 INFO:teuthology.orchestra.run.vm04.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-10T12:57:09.486 INFO:teuthology.orchestra.run.vm04.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T12:57:09.487 INFO:teuthology.orchestra.run.vm04.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Getting image source signatures 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T12:57:10.799 INFO:teuthology.orchestra.run.vm04.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-10T12:57:10.938 INFO:teuthology.orchestra.run.vm04.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T12:57:10.938 INFO:teuthology.orchestra.run.vm04.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T12:57:10.938 INFO:teuthology.orchestra.run.vm04.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T12:57:11.018 INFO:teuthology.orchestra.run.vm04.stdout:stat: stdout 167 167 2026-03-10T12:57:11.018 INFO:teuthology.orchestra.run.vm04.stdout:Creating initial keys... 2026-03-10T12:57:11.115 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph-authtool: stdout AQAnFbBpyPtjBRAAUmokEpBoOv6OqfWNtnxfCg== 2026-03-10T12:57:11.242 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph-authtool: stdout AQAnFbBpgQPLCxAAjvWk4zwgDUbtB110BYylAw== 2026-03-10T12:57:11.333 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph-authtool: stdout AQAnFbBpduqPEhAANvzA918G9+ReFJ8MnDIWoQ== 2026-03-10T12:57:11.334 INFO:teuthology.orchestra.run.vm04.stdout:Creating initial monmap... 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:monmaptool for a [v2:192.168.123.104:3300,v1:192.168.123.104:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:setting min_mon_release = quincy 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: set fsid to 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:11.425 INFO:teuthology.orchestra.run.vm04.stdout:Creating mon... 2026-03-10T12:57:11.601 INFO:teuthology.orchestra.run.vm04.stdout:create mon.a on 2026-03-10T12:57:11.768 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-10T12:57:11.901 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T12:57:12.037 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target → /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target. 2026-03-10T12:57:12.037 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target → /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target. 2026-03-10T12:57:12.195 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a 2026-03-10T12:57:12.195 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Failed to reset failed state of unit ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service: Unit ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service not loaded. 2026-03-10T12:57:12.333 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target.wants/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service → /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@.service. 2026-03-10T12:57:12.525 INFO:teuthology.orchestra.run.vm04.stdout:firewalld does not appear to be present 2026-03-10T12:57:12.526 INFO:teuthology.orchestra.run.vm04.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T12:57:12.526 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mon to start... 2026-03-10T12:57:12.526 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mon... 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout id: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout services: 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.11591s) 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout data: 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T12:57:12.727 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T12:57:12.728 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T12:57:12.728 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.728 INFO:teuthology.orchestra.run.vm04.stdout:mon is available 2026-03-10T12:57:12.728 INFO:teuthology.orchestra.run.vm04.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T12:57:12.946 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout fsid = 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.104:3300,v1:192.168.123.104:6789] 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T12:57:12.947 INFO:teuthology.orchestra.run.vm04.stdout:Generating new minimal ceph.conf... 2026-03-10T12:57:13.125 INFO:teuthology.orchestra.run.vm04.stdout:Restarting the monitor... 2026-03-10T12:57:13.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a[50338]: 2026-03-10T12:57:13.214+0000 7f6855b51640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-10T12:57:13.668 INFO:teuthology.orchestra.run.vm04.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-10T12:57:13.705 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50538]: 2026-03-10 12:57:13.449909044 +0000 UTC m=+0.266368299 container died abc1fb8f30b68eb4642ddc6ff0cd3e09d650ed499ba6286177ddf77e72da3552 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T12:57:13.705 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50538]: 2026-03-10 12:57:13.461948839 +0000 UTC m=+0.278408094 container remove abc1fb8f30b68eb4642ddc6ff0cd3e09d650ed499ba6286177ddf77e72da3552 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, ceph=True, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T12:57:13.705 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 bash[50538]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a 2026-03-10T12:57:13.705 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service: Deactivated successfully. 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 systemd[1]: Stopped Ceph mon.a for 74d0e8a6-1c80-11f1-9542-4d7926099318. 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 systemd[1]: Starting Ceph mon.a for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50606]: 2026-03-10 12:57:13.620456592 +0000 UTC m=+0.017353449 container create 7ec89c0a4b8acf9da20a8b747f28f8db1b3575be06ef5b9a5a05ced5fdd775a1 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.schema-version=1.0) 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50606]: 2026-03-10 12:57:13.654377556 +0000 UTC m=+0.051274424 container init 7ec89c0a4b8acf9da20a8b747f28f8db1b3575be06ef5b9a5a05ced5fdd775a1 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50606]: 2026-03-10 12:57:13.656801517 +0000 UTC m=+0.053698374 container start 7ec89c0a4b8acf9da20a8b747f28f8db1b3575be06ef5b9a5a05ced5fdd775a1 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , ceph=True, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 bash[50606]: 7ec89c0a4b8acf9da20a8b747f28f8db1b3575be06ef5b9a5a05ced5fdd775a1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 podman[50606]: 2026-03-10 12:57:13.613540456 +0000 UTC m=+0.010437323 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 systemd[1]: Started Ceph mon.a for 74d0e8a6-1c80-11f1-9542-4d7926099318. 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: set uid:gid to 167:167 (ceph:ceph) 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable), process ceph-mon, pid 2 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: pidfile_write: ignore empty --pid-file 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: load: jerasure load: lrc 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: RocksDB version: 7.9.2 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Git sha 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Compile date 2026-02-25 18:11:04 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: DB SUMMARY 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: DB Session ID: MBO8XYFGYDETOFZWR8CB 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: CURRENT file: CURRENT 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: IDENTITY file: IDENTITY 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 75535 ; 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.error_if_exists: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.create_if_missing: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.paranoid_checks: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.flush_verify_memtable_count: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.env: 0x55b09d491dc0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.fs: PosixFileSystem 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.info_log: 0x55b09e89b880 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_file_opening_threads: 16 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.statistics: (nil) 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.use_fsync: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_log_file_size: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.log_file_time_to_roll: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.keep_log_file_num: 1000 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.recycle_log_file_num: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_fallocate: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_mmap_reads: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_mmap_writes: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.use_direct_reads: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.create_missing_column_families: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.db_log_dir: 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.wal_dir: 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.table_cache_numshardbits: 6 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.WAL_ttl_seconds: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.WAL_size_limit_MB: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.is_fd_close_on_exec: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.advise_random_on_open: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.db_write_buffer_size: 0 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.write_buffer_manager: 0x55b09e89f900 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-10T12:57:13.706 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.use_adaptive_mutex: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.rate_limiter: (nil) 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.wal_recovery_mode: 2 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enable_thread_tracking: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enable_pipelined_write: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.unordered_write: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.row_cache: None 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.wal_filter: None 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_ingest_behind: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.two_write_queues: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.manual_wal_flush: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.wal_compression: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.atomic_flush: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.persist_stats_to_disk: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.write_dbid_to_manifest: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.log_readahead_size: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.best_efforts_recovery: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.allow_data_in_errors: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.db_host_id: __hostname__ 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enforce_single_del_contracts: true 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_background_jobs: 2 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_background_compactions: -1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_subcompactions: 1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.delayed_write_rate : 16777216 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_total_wal_size: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.stats_dump_period_sec: 600 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.stats_persist_period_sec: 600 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_open_files: -1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bytes_per_sync: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.wal_bytes_per_sync: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.strict_bytes_per_sync: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_readahead_size: 0 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_background_flushes: -1 2026-03-10T12:57:13.707 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Compression algorithms supported: 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kZSTD supported: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kXpressCompression supported: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kBZip2Compression supported: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kLZ4Compression supported: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kZlibCompression supported: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kLZ4HCCompression supported: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: kSnappyCompression supported: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Fast CRC32 supported: Supported on x86 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: DMutex implementation: pthread_mutex_t 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.merge_operator: 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_filter: None 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_filter_factory: None 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.sst_partitioner_factory: None 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.memtable_factory: SkipListFactory 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.table_factory: BlockBasedTable 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55b09e89b4e0) 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: cache_index_and_filter_blocks: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: pin_top_level_index_and_filter: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: index_type: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: data_block_index_type: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: index_shortening: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: data_block_hash_table_util_ratio: 0.750000 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: checksum: 4 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: no_block_cache: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_cache: 0x55b09e8be9b0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_cache_name: BinnedLRUCache 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_cache_options: 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: capacity : 536870912 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: num_shard_bits : 4 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: strict_capacity_limit : 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: high_pri_pool_ratio: 0.000 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_cache_compressed: (nil) 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: persistent_cache: (nil) 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_size: 4096 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_size_deviation: 10 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_restart_interval: 16 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: index_block_restart_interval: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: metadata_block_size: 4096 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: partition_filters: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: use_delta_encoding: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: filter_policy: bloomfilter 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: whole_key_filtering: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: verify_compression: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: read_amp_bytes_per_bit: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: format_version: 5 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: enable_index_compression: 1 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: block_align: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: max_auto_readahead_size: 262144 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: prepopulate_block_cache: 0 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: initial_auto_readahead_size: 8192 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout: num_file_reads_for_auto_readahead: 2 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.write_buffer_size: 33554432 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_write_buffer_number: 2 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression: NoCompression 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression: Disabled 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.prefix_extractor: nullptr 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-10T12:57:13.708 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.num_levels: 7 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.window_bits: -14 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.level: 32767 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.strategy: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.enabled: false 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.target_file_size_base: 67108864 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.target_file_size_multiplier: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.arena_block_size: 1048576 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.disable_auto_compactions: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.inplace_update_support: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.inplace_update_num_locks: 10000 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.memtable_huge_page_size: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.bloom_locality: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.max_successive_merges: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.optimize_filters_for_hits: 0 2026-03-10T12:57:13.709 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.paranoid_file_checks: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.force_consistency_checks: 1 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.report_bg_io_stats: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.ttl: 2592000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.periodic_compaction_seconds: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enable_blob_files: false 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.min_blob_size: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_file_size: 268435456 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_compression_type: NoCompression 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.enable_blob_garbage_collection: false 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.blob_file_starting_level: 0 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: f1beb786-0372-4650-a66b-94ccda01136a 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773147433681759, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773147433683275, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72616, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 225, "table_properties": {"data_size": 70895, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9705, "raw_average_key_size": 49, "raw_value_size": 65374, "raw_average_value_size": 333, "num_data_blocks": 8, "num_entries": 196, "num_filter_entries": 196, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1773147433, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "f1beb786-0372-4650-a66b-94ccda01136a", "db_session_id": "MBO8XYFGYDETOFZWR8CB", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: EVENT_LOG_v1 {"time_micros": 1773147433683343, "job": 1, "event": "recovery_finished"} 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55b09e8c0e00 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: rocksdb: DB pointer 0x55b09e8d0000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: starting mon.a rank 0 at public addrs [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] at bind addrs [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???) e1 preinit fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).mds e1 new map 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).mds e1 print_map 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: e1 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: btime 2026-03-10T12:57:12:584673+0000 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: enable_multiple, ever_enabled_multiple: 1,1 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: legacy client fscid: -1 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout: No filesystems configured 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-10T12:57:13.710 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-10T12:57:13.890 INFO:teuthology.orchestra.run.vm04.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T12:57:13.891 INFO:teuthology.orchestra.run.vm04.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:13.891 INFO:teuthology.orchestra.run.vm04.stdout:Creating mgr... 2026-03-10T12:57:13.892 INFO:teuthology.orchestra.run.vm04.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T12:57:13.892 INFO:teuthology.orchestra.run.vm04.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mon.a is new leader, mons a in quorum (ranks 0) 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: monmap epoch 1 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: last_changed 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: min_mon_release 19 (squid) 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: election_strategy: 1 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: fsmap 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: osdmap e1: 0 total, 0 up, 0 in 2026-03-10T12:57:13.962 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:13 vm04 ceph-mon[50620]: mgrmap e1: no daemons active 2026-03-10T12:57:14.047 INFO:teuthology.orchestra.run.vm04.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a 2026-03-10T12:57:14.047 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Failed to reset failed state of unit ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a.service: Unit ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a.service not loaded. 2026-03-10T12:57:14.184 INFO:teuthology.orchestra.run.vm04.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318.target.wants/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a.service → /etc/systemd/system/ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@.service. 2026-03-10T12:57:14.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 systemd[1]: Starting Ceph mgr.a for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T12:57:14.588 INFO:teuthology.orchestra.run.vm04.stdout:firewalld does not appear to be present 2026-03-10T12:57:14.588 INFO:teuthology.orchestra.run.vm04.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T12:57:14.588 INFO:teuthology.orchestra.run.vm04.stdout:firewalld does not appear to be present 2026-03-10T12:57:14.588 INFO:teuthology.orchestra.run.vm04.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-10T12:57:14.588 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mgr to start... 2026-03-10T12:57:14.589 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mgr... 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 podman[50819]: 2026-03-10 12:57:14.301458526 +0000 UTC m=+0.011857234 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 podman[50819]: 2026-03-10 12:57:14.507691482 +0000 UTC m=+0.218090190 container create d39e4ab96c22ca50cdf57705b14032d200bd3a60746df3f31095d5426882852b (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a, CEPH_REF=squid, org.label-schema.schema-version=1.0, ceph=True, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 podman[50819]: 2026-03-10 12:57:14.57601327 +0000 UTC m=+0.286411958 container init d39e4ab96c22ca50cdf57705b14032d200bd3a60746df3f31095d5426882852b (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, OSD_FLAVOR=default, io.buildah.version=1.41.3) 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 podman[50819]: 2026-03-10 12:57:14.578916369 +0000 UTC m=+0.289315067 container start d39e4ab96c22ca50cdf57705b14032d200bd3a60746df3f31095d5426882852b (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid) 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 bash[50819]: d39e4ab96c22ca50cdf57705b14032d200bd3a60746df3f31095d5426882852b 2026-03-10T12:57:14.657 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 systemd[1]: Started Ceph mgr.a for 74d0e8a6-1c80-11f1-9542-4d7926099318. 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsid": "74d0e8a6-1c80-11f1-9542-4d7926099318", 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:57:14.811 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:57:12:584673+0000", 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:57:14.813 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:57:12.585228+0000", 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:14.814 INFO:teuthology.orchestra.run.vm04.stdout:mgr not available, waiting (1/15)... 2026-03-10T12:57:15.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/412785702' entity='client.admin' 2026-03-10T12:57:15.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/627888546' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:57:15.127 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:14.691+0000 7f7a68d0b140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:57:15.127 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:14 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:14.750+0000 7f7a68d0b140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:57:15.546 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:15.197+0000 7f7a68d0b140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:15.543+0000 7f7a68d0b140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: from numpy import show_config as show_numpy_config 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:15.636+0000 7f7a68d0b140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:15.675+0000 7f7a68d0b140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:57:15.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:15 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:15.751+0000 7f7a68d0b140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:57:16.546 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.264+0000 7f7a68d0b140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:57:16.546 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.383+0000 7f7a68d0b140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:16.546 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.422+0000 7f7a68d0b140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:57:16.546 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.459+0000 7f7a68d0b140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:16.547 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.503+0000 7f7a68d0b140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:57:16.852 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.543+0000 7f7a68d0b140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:57:16.852 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.742+0000 7f7a68d0b140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:57:16.852 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:16 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:16.793+0000 7f7a68d0b140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:17.112 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3980466658' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:57:17.112 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.061+0000 7f7a68d0b140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsid": "74d0e8a6-1c80-11f1-9542-4d7926099318", 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:57:17.117 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:57:12:584673+0000", 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:57:12.585228+0000", 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:17.118 INFO:teuthology.orchestra.run.vm04.stdout:mgr not available, waiting (2/15)... 2026-03-10T12:57:17.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.370+0000 7f7a68d0b140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:57:17.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.407+0000 7f7a68d0b140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:57:17.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.455+0000 7f7a68d0b140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:57:17.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.540+0000 7f7a68d0b140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:57:17.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.578+0000 7f7a68d0b140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:57:17.927 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.663+0000 7f7a68d0b140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:57:17.927 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.778+0000 7f7a68d0b140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: Activating manager daemon a 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: mgrmap e2: a(active, starting, since 0.0641727s) 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: Manager daemon a is now available 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:18 vm04 ceph-mon[50620]: from='mgr.14100 192.168.123.104:0/2239591828' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.924+0000 7f7a68d0b140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:57:18.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:17.962+0000 7f7a68d0b140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:57:19.399 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsid": "74d0e8a6-1c80-11f1-9542-4d7926099318", 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 0 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T12:57:19.402 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "a" 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T12:57:12:584673+0000", 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ], 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T12:57:12.585228+0000", 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout }, 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:19.403 INFO:teuthology.orchestra.run.vm04.stdout:mgr is available 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout fsid = 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.104:3300,v1:192.168.123.104:6789] 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T12:57:19.706 INFO:teuthology.orchestra.run.vm04.stdout:Enabling cephadm module... 2026-03-10T12:57:20.035 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-mon[50620]: mgrmap e3: a(active, since 1.06816s) 2026-03-10T12:57:20.035 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2921397051' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-10T12:57:20.035 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/899118872' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-10T12:57:20.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/899118872' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-10T12:57:20.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2584114019' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-10T12:57:21.037 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: ignoring --setuser ceph since I am not root 2026-03-10T12:57:21.038 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: ignoring --setgroup ceph since I am not root 2026-03-10T12:57:21.038 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:20.778+0000 7fde4b043140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:57:21.038 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:20 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:20.823+0000 7fde4b043140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:57:21.107 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:21.107 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for the mgr to restart... 2026-03-10T12:57:21.108 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mgr epoch 4... 2026-03-10T12:57:21.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:21.274+0000 7fde4b043140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2584114019' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-mon[50620]: mgrmap e4: a(active, since 2s) 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2013055388' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:21.612+0000 7fde4b043140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: from numpy import show_config as show_numpy_config 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:21.703+0000 7fde4b043140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:21.742+0000 7fde4b043140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:57:21.877 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:21.814+0000 7fde4b043140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:57:22.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.346+0000 7fde4b043140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:57:22.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.471+0000 7fde4b043140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:22.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.514+0000 7fde4b043140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:57:22.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.551+0000 7fde4b043140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:22.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.594+0000 7fde4b043140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:57:23.100 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.633+0000 7fde4b043140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:57:23.100 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.809+0000 7fde4b043140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:57:23.100 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:22 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:22.861+0000 7fde4b043140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:23.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.097+0000 7fde4b043140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:57:23.687 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.392+0000 7fde4b043140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:57:23.687 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.429+0000 7fde4b043140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:57:23.687 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.472+0000 7fde4b043140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:57:23.687 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.560+0000 7fde4b043140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:57:23.687 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.600+0000 7fde4b043140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:57:23.974 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.684+0000 7fde4b043140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:57:23.974 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.799+0000 7fde4b043140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:23.974 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.933+0000 7fde4b043140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: Active manager daemon a restarted 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: Activating manager daemon a 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: osdmap e2: 0 total, 0 up, 0 in 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: mgrmap e5: a(active, starting, since 0.00736391s) 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: Manager daemon a is now available 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:24 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:57:24.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:23.971+0000 7fde4b043140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:57:25.027 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:25.028 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-10T12:57:25.028 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T12:57:25.028 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:25.028 INFO:teuthology.orchestra.run.vm04.stdout:mgr epoch 4 is available 2026-03-10T12:57:25.028 INFO:teuthology.orchestra.run.vm04.stdout:Setting orchestrator backend to cephadm... 2026-03-10T12:57:25.308 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-mon[50620]: Found migration_current of "None". Setting to last migration. 2026-03-10T12:57:25.308 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:57:25.308 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:25.308 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:25.308 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-mon[50620]: mgrmap e6: a(active, since 1.0101s) 2026-03-10T12:57:25.632 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T12:57:25.632 INFO:teuthology.orchestra.run.vm04.stdout:Generating ssh key... 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Generating public/private ed25519 key pair. 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Your identification has been saved in /tmp/tmpcaidrgl5/key 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Your public key has been saved in /tmp/tmpcaidrgl5/key.pub 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: The key fingerprint is: 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: SHA256:OswFE12XUb2P3xYITQobGlRrknKFCqiF87Ky/kV1phY ceph-74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: The key's randomart image is: 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: +--[ED25519 256]--+ 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: | .. ooo+. o+.. | 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |o... ++o.... .| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |.+ ..+Eo=+ + .| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |o . .++B. o . . | 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: | o . oS . ...| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |o .o.o ....| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |.. .= .o| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |. . . +| 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: |.... . | 2026-03-10T12:57:26.115 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: +----[SHA256]-----+ 2026-03-10T12:57:26.157 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:26.157 INFO:teuthology.orchestra.run.vm04.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T12:57:26.158 INFO:teuthology.orchestra.run.vm04.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T12:57:26.158 INFO:teuthology.orchestra.run.vm04.stdout:Adding host vm04... 2026-03-10T12:57:26.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:25] ENGINE Bus STARTING 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:25] ENGINE Serving on http://192.168.123.104:8765 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:26.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:26 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:25] ENGINE Serving on https://192.168.123.104:7150 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:25] ENGINE Bus STARTED 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:25] ENGINE Client ('192.168.123.104', 34760) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: Generating ssh key... 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "addr": "192.168.123.104", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:27.584 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:27 vm04 ceph-mon[50620]: mgrmap e7: a(active, since 2s) 2026-03-10T12:57:27.984 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout Added host 'vm04' with addr '192.168.123.104' 2026-03-10T12:57:27.984 INFO:teuthology.orchestra.run.vm04.stdout:Deploying unmanaged mon service... 2026-03-10T12:57:28.264 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T12:57:28.264 INFO:teuthology.orchestra.run.vm04.stdout:Deploying unmanaged mgr service... 2026-03-10T12:57:28.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:28 vm04 ceph-mon[50620]: Deploying cephadm binary to vm04 2026-03-10T12:57:28.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:28 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:28.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:28 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:28.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:28 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:29.451 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: Added host vm04 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: Saving service mon spec with placement count:5 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: Saving service mgr spec with placement count:2 2026-03-10T12:57:29.764 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:29 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:30.086 INFO:teuthology.orchestra.run.vm04.stdout:Enabling the dashboard module... 2026-03-10T12:57:30.994 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:30.994 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/954057903' entity='client.admin' 2026-03-10T12:57:30.994 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:30.995 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/4112092680' entity='client.admin' 2026-03-10T12:57:30.995 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/653986964' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T12:57:30.995 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:30 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:31.255 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: ignoring --setuser ceph since I am not root 2026-03-10T12:57:31.255 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: ignoring --setgroup ceph since I am not root 2026-03-10T12:57:31.255 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:31.228+0000 7f2d8e4eb140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T12:57:31.428 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:31.429 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for the mgr to restart... 2026-03-10T12:57:31.429 INFO:teuthology.orchestra.run.vm04.stdout:Waiting for mgr epoch 8... 2026-03-10T12:57:31.508 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:31.273+0000 7f2d8e4eb140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:57:31.798 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:31.713+0000 7f2d8e4eb140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:57:32.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-mon[50620]: from='mgr.14118 192.168.123.104:0/247457648' entity='mgr.a' 2026-03-10T12:57:32.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/653986964' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T12:57:32.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-mon[50620]: mgrmap e8: a(active, since 7s) 2026-03-10T12:57:32.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:31 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/4009447619' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T12:57:32.127 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.044+0000 7f2d8e4eb140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: from numpy import show_config as show_numpy_config 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.135+0000 7f2d8e4eb140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.170+0000 7f2d8e4eb140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:57:32.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.241+0000 7f2d8e4eb140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:57:32.993 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.728+0000 7f2d8e4eb140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:57:32.993 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.840+0000 7f2d8e4eb140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:32.993 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.879+0000 7f2d8e4eb140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:57:32.993 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.914+0000 7f2d8e4eb140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:32.993 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.954+0000 7f2d8e4eb140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:57:33.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:32 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:32.990+0000 7f2d8e4eb140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:57:33.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.160+0000 7f2d8e4eb140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:57:33.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.210+0000 7f2d8e4eb140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:57:33.711 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.431+0000 7f2d8e4eb140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:57:33.987 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.708+0000 7f2d8e4eb140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:57:33.987 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.747+0000 7f2d8e4eb140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:57:33.988 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.792+0000 7f2d8e4eb140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:57:33.988 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.869+0000 7f2d8e4eb140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:57:33.988 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.906+0000 7f2d8e4eb140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:57:34.270 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:33 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:33.984+0000 7f2d8e4eb140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:57:34.270 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:34.100+0000 7f2d8e4eb140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:57:34.270 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:34.230+0000 7f2d8e4eb140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:57:34.626 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:34.267+0000 7f2d8e4eb140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: Active manager daemon a restarted 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: Activating manager daemon a 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: mgrmap e9: a(active, starting, since 0.0116296s) 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: Manager daemon a is now available 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-10T12:57:34.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout { 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout } 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:mgr epoch 8 is available 2026-03-10T12:57:35.371 INFO:teuthology.orchestra.run.vm04.stdout:Generating a dashboard self-signed certificate... 2026-03-10T12:57:35.744 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T12:57:35.744 INFO:teuthology.orchestra.run.vm04.stdout:Creating initial admin user... 2026-03-10T12:57:35.959 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:36.188 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$EO1rea.sd1U83v5EM6HUNuRDKnuZOINIg2aqcLY2YAC7xRZb6GzIe", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773147456, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T12:57:36.189 INFO:teuthology.orchestra.run.vm04.stdout:Fetching dashboard port number... 2026-03-10T12:57:36.216 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:35 vm04 ceph-mon[50620]: mgrmap e10: a(active, since 1.01483s) 2026-03-10T12:57:36.216 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:36.216 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:36.216 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:36.466 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T12:57:36.466 INFO:teuthology.orchestra.run.vm04.stdout:firewalld does not appear to be present 2026-03-10T12:57:36.466 INFO:teuthology.orchestra.run.vm04.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout:Ceph Dashboard is now available at: 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout: URL: https://vm04.local:8443/ 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout: User: admin 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout: Password: kk22c0by5j 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.468 INFO:teuthology.orchestra.run.vm04.stdout:Saving cluster configuration to /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config directory 2026-03-10T12:57:36.786 INFO:teuthology.orchestra.run.vm04.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: ceph telemetry on 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout:For more information see: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:36.787 INFO:teuthology.orchestra.run.vm04.stdout:Bootstrap complete. 2026-03-10T12:57:36.821 INFO:tasks.cephadm:Fetching config... 2026-03-10T12:57:36.821 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:57:36.821 DEBUG:teuthology.orchestra.run.vm04:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T12:57:36.868 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T12:57:36.869 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:57:36.869 DEBUG:teuthology.orchestra.run.vm04:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T12:57:36.896 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T12:57:36.897 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:57:36.897 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/keyring of=/dev/stdout 2026-03-10T12:57:36.964 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T12:57:36.964 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:57:36.964 DEBUG:teuthology.orchestra.run.vm04:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T12:57:37.021 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T12:57:37.021 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T12:57:37.133 INFO:teuthology.orchestra.run.vm04.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:37.147 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T12:57:37.182 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:37.193 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T12:57:37.230 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOoJgFF0AuUBoyxZqZlMUZuWpmVBuB6NiCuwlOTv+N+Q ceph-74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:37.241 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:35] ENGINE Bus STARTING 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:35] ENGINE Serving on http://192.168.123.104:8765 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:35] ENGINE Serving on https://192.168.123.104:7150 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:35] ENGINE Bus STARTED 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: [10/Mar/2026:12:57:35] ENGINE Client ('192.168.123.104', 52764) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1751004652' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T12:57:37.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:37 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/650004279' entity='client.admin' 2026-03-10T12:57:37.437 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:37.813 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T12:57:37.813 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T12:57:38.024 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:38.316 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:38 vm04 ceph-mon[50620]: mgrmap e11: a(active, since 2s) 2026-03-10T12:57:38.316 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:38 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1008010398' entity='client.admin' 2026-03-10T12:57:38.377 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm06 2026-03-10T12:57:38.377 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:57:38.377 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.conf 2026-03-10T12:57:38.393 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:57:38.394 DEBUG:teuthology.orchestra.run.vm06:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:38.450 INFO:tasks.cephadm:Adding host vm06 to orchestrator... 2026-03-10T12:57:38.450 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch host add vm06 2026-03-10T12:57:38.659 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:39.304 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:39.304 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:39.304 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm06", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:39.304 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:39.305 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:39.305 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:57:39.305 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:39.305 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:57:39.305 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:39 vm04 ceph-mon[50620]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:57:40.526 INFO:teuthology.orchestra.run.vm04.stdout:Added host 'vm06' with addr '192.168.123.106' 2026-03-10T12:57:40.584 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch host ls --format=json 2026-03-10T12:57:40.771 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: Deploying cephadm binary to vm06 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.client.admin.keyring 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:40.794 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:40.999 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:40.999 INFO:teuthology.orchestra.run.vm04.stdout:[{"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}] 2026-03-10T12:57:41.050 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-10T12:57:41.050 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:57:41.050 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-10T12:57:41.067 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:57:41.067 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:41.124 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-10T12:57:41.124 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch host add vm09 2026-03-10T12:57:41.289 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:42.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:41 vm04 ceph-mon[50620]: Added host vm06 2026-03-10T12:57:42.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:42.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:41 vm04 ceph-mon[50620]: mgrmap e12: a(active, since 6s) 2026-03-10T12:57:42.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:42.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:41 vm04 ceph-mon[50620]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:57:43.051 INFO:teuthology.orchestra.run.vm04.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-10T12:57:43.051 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:42 vm04 ceph-mon[50620]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:43.051 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:43.051 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:42 vm04 ceph-mon[50620]: Deploying cephadm binary to vm09 2026-03-10T12:57:43.102 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch host ls --format=json 2026-03-10T12:57:43.300 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:43.555 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:57:43.555 INFO:teuthology.orchestra.run.vm04.stdout:[{"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-10T12:57:43.619 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T12:57:43.620 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd crush tunables default 2026-03-10T12:57:43.832 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: Added host vm09 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:57:44.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:45.056 INFO:teuthology.orchestra.run.vm04.stderr:adjusted tunables profile to default 2026-03-10T12:57:45.110 INFO:tasks.cephadm:Adding mon.a on vm04 2026-03-10T12:57:45.110 INFO:tasks.cephadm:Adding mon.b on vm06 2026-03-10T12:57:45.110 INFO:tasks.cephadm:Adding mon.c on vm09 2026-03-10T12:57:45.110 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch apply mon '3;vm04:192.168.123.104=a;vm06:192.168.123.106=b;vm09:192.168.123.109=c' 2026-03-10T12:57:45.320 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T12:57:45.373 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.client.admin.keyring 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2158836614' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T12:57:45.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:45 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:45.664 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-10T12:57:45.720 DEBUG:teuthology.orchestra.run.vm06:mon.b> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.b.service 2026-03-10T12:57:45.722 DEBUG:teuthology.orchestra.run.vm09:mon.c> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.c.service 2026-03-10T12:57:45.724 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T12:57:45.724 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph mon dump -f json 2026-03-10T12:57:45.999 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T12:57:46.057 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T12:57:46.353 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:57:46.353 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","modified":"2026-03-10T12:57:11.405959Z","created":"2026-03-10T12:57:11.405959Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T12:57:46.353 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T12:57:46.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:46 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2158836614' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T12:57:46.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:46 vm04 ceph-mon[50620]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:57:46.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:46 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:46.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:46 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm04:192.168.123.104=a;vm06:192.168.123.106=b;vm09:192.168.123.109=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: Saving service mon spec with placement vm04:192.168.123.104=a;vm06:192.168.123.106=b;vm09:192.168.123.109=c;count:3 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/2119485967' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:47.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:57:47.430 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T12:57:47.430 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph mon dump -f json 2026-03-10T12:57:47.716 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:57:48.045 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:57:48.045 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","modified":"2026-03-10T12:57:11.405959Z","created":"2026-03-10T12:57:11.405959Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T12:57:48.046 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:48.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:48 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/1768396866' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:57:49.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:48 vm09 ceph-mon[53016]: mon.c@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T12:57:49.220 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-10T12:57:49.220 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph mon dump -f json 2026-03-10T12:57:49.401 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:57:50.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:57:50 vm06 ceph-mon[56046]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: Deploying daemon mon.b on vm06 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: mon.a calling monitor election 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: mon.c calling monitor election 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: monmap epoch 2 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: last_changed 2026-03-10T12:57:48.949111+0000 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: min_mon_release 19 (squid) 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: election_strategy: 1 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: fsmap 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: mgrmap e12: a(active, since 19s) 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: overall HEALTH_OK 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.342 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.343 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.343 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: Deploying daemon mon.b on vm06 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: mon.a calling monitor election 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: mon.c calling monitor election 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: monmap epoch 2 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: last_changed 2026-03-10T12:57:48.949111+0000 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: min_mon_release 19 (squid) 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: election_strategy: 1 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: fsmap 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: mgrmap e12: a(active, since 19s) 2026-03-10T12:57:54.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: overall HEALTH_OK 2026-03-10T12:57:54.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:54.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:55.377 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:57:54 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:57:54.947+0000 7f2d5a854640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: mon.c calling monitor election 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: mon.a calling monitor election 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: monmap epoch 3 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: min_mon_release 19 (squid) 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: election_strategy: 1 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: fsmap 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: mgrmap e12: a(active, since 25s) 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: overall HEALTH_OK 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:59.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:57:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: mon.c calling monitor election 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: mon.a calling monitor election 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: monmap epoch 3 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: min_mon_release 19 (squid) 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: election_strategy: 1 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: fsmap 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: mgrmap e12: a(active, since 25s) 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: overall HEALTH_OK 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:57:59.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:57:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:00.073 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:58:00.073 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":3,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","modified":"2026-03-10T12:57:54.377522Z","created":"2026-03-10T12:57:11.405959Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T12:58:00.073 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 3 2026-03-10T12:58:00.157 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T12:58:00.157 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph config generate-minimal-conf 2026-03-10T12:58:00.336 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:58:00.588 INFO:teuthology.orchestra.run.vm04.stdout:# minimal ceph.conf for 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:00.588 INFO:teuthology.orchestra.run.vm04.stdout:[global] 2026-03-10T12:58:00.588 INFO:teuthology.orchestra.run.vm04.stdout: fsid = 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:00.588 INFO:teuthology.orchestra.run.vm04.stdout: mon_host = [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T12:58:00.676 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T12:58:00.676 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:58:00.676 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T12:58:00.703 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:58:00.703 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.767 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Reconfiguring daemon mon.a on vm04 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Reconfiguring mon.b (monmap changed)... 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: Reconfiguring daemon mon.b on vm06 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/697935088' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.768 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:00 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/300096791' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.771 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:58:00.771 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T12:58:00.800 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:58:00.800 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Reconfiguring daemon mon.a on vm04 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Reconfiguring mon.b (monmap changed)... 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: Reconfiguring daemon mon.b on vm06 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='client.? 192.168.123.109:0/697935088' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:00.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:00 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/300096791' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:00.871 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:58:00.871 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T12:58:00.896 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:58:00.896 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:58:00.961 INFO:tasks.cephadm:Adding mgr.a on vm04 2026-03-10T12:58:00.961 INFO:tasks.cephadm:Adding mgr.b on vm06 2026-03-10T12:58:00.961 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch apply mgr '2;vm04=a;vm06=b' 2026-03-10T12:58:01.167 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:58:01.393 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mgr update... 2026-03-10T12:58:01.468 DEBUG:teuthology.orchestra.run.vm06:mgr.b> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.b.service 2026-03-10T12:58:01.469 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T12:58:01.469 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-10T12:58:01.469 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T12:58:01.485 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:58:01.486 DEBUG:teuthology.orchestra.run.vm04:> ls /dev/[sv]d? 2026-03-10T12:58:01.544 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vda 2026-03-10T12:58:01.544 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdb 2026-03-10T12:58:01.544 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdc 2026-03-10T12:58:01.544 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vdd 2026-03-10T12:58:01.544 INFO:teuthology.orchestra.run.vm04.stdout:/dev/vde 2026-03-10T12:58:01.544 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T12:58:01.544 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T12:58:01.544 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdb 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdb 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-10 12:57:38.888942325 +0000 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-10 12:54:13.598576611 +0000 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-10 12:54:13.598576611 +0000 2026-03-10T12:58:01.603 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-10 12:51:24.209000000 +0000 2026-03-10T12:58:01.603 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T12:58:01.650 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Deploying daemon mon.b on vm06 2026-03-10T12:58:01.650 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:01.650 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.a calling monitor election 2026-03-10T12:58:01.650 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.c calling monitor election 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: monmap epoch 2 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: last_changed 2026-03-10T12:57:48.949111+0000 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: min_mon_release 19 (squid) 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: election_strategy: 1 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: fsmap 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mgrmap e12: a(active, since 19s) 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: overall HEALTH_OK 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.c calling monitor election 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.a calling monitor election 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-10T12:58:01.651 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: monmap epoch 3 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: min_mon_release 19 (squid) 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: election_strategy: 1 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: fsmap 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: mgrmap e12: a(active, since 25s) 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: overall HEALTH_OK 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Reconfiguring mon.a (unknown last config time)... 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Reconfiguring daemon mon.a on vm04 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Reconfiguring mon.b (monmap changed)... 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: Reconfiguring daemon mon.b on vm06 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='client.? 192.168.123.109:0/697935088' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:01.652 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:01 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/300096791' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:01.669 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-10T12:58:01.669 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-10T12:58:01.669 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000201778 s, 2.5 MB/s 2026-03-10T12:58:01.670 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T12:58:01.729 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdc 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdc 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 242 Links: 1 Device type: fc,20 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-10 12:57:38.921942407 +0000 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-10 12:54:13.583576580 +0000 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-10 12:54:13.583576580 +0000 2026-03-10T12:58:01.788 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-10 12:51:24.226000000 +0000 2026-03-10T12:58:01.788 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T12:58:01.853 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-10T12:58:01.853 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-10T12:58:01.853 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000217819 s, 2.4 MB/s 2026-03-10T12:58:01.854 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T12:58:01.910 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vdd 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vdd 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-10 12:57:38.960942505 +0000 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-10 12:54:13.604576623 +0000 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-10 12:54:13.604576623 +0000 2026-03-10T12:58:01.968 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-10 12:51:24.229000000 +0000 2026-03-10T12:58:01.969 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T12:58:02.035 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-10T12:58:02.035 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-10T12:58:02.035 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000167094 s, 3.1 MB/s 2026-03-10T12:58:02.037 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T12:58:02.093 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vde 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vde 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-10 12:57:38.994942590 +0000 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-10 12:54:13.584576582 +0000 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-10 12:54:13.584576582 +0000 2026-03-10T12:58:02.150 INFO:teuthology.orchestra.run.vm04.stdout: Birth: 2026-03-10 12:51:24.234000000 +0000 2026-03-10T12:58:02.151 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T12:58:02.221 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-10T12:58:02.221 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-10T12:58:02.221 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000196267 s, 2.6 MB/s 2026-03-10T12:58:02.223 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T12:58:02.285 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T12:58:02.285 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T12:58:02.308 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:02.306+0000 7fa3a72e5140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-10T12:58:02.334 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:58:02.334 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-10T12:58:02.452 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-10T12:58:02.452 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-10T12:58:02.452 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-10T12:58:02.452 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-10T12:58:02.452 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-10T12:58:02.452 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T12:58:02.452 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T12:58:02.452 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-10T12:58:02.517 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 223 Links: 1 Device type: fc,10 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 12:57:42.952699654 +0000 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 12:54:13.827429096 +0000 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 12:54:13.827429096 +0000 2026-03-10T12:58:02.518 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 12:50:53.224000000 +0000 2026-03-10T12:58:02.518 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T12:58:02.559 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T12:58:02.559 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T12:58:02.559 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000328394 s, 1.6 MB/s 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mon.b calling monitor election 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm04=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: Saving service mgr spec with placement vm04=a;vm06=b;count:2 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mon.b calling monitor election 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mon.a calling monitor election 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mon.c calling monitor election 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: monmap epoch 3 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: min_mon_release 19 (squid) 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: election_strategy: 1 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: fsmap 2026-03-10T12:58:02.560 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: mgrmap e12: a(active, since 27s) 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: overall HEALTH_OK 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: Deploying daemon mgr.b on vm06 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:02.561 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:02.561 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T12:58:02.630 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 12:57:42.979699657 +0000 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 12:54:13.834429097 +0000 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 12:54:13.834429097 +0000 2026-03-10T12:58:02.723 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 12:50:53.244000000 +0000 2026-03-10T12:58:02.723 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T12:58:02.772 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T12:58:02.772 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T12:58:02.772 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000173174 s, 3.0 MB/s 2026-03-10T12:58:02.773 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T12:58:02.819 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-10T12:58:02.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mon.b calling monitor election 2026-03-10T12:58:02.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm04=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:02.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: Saving service mgr spec with placement vm04=a;vm06=b;count:2 2026-03-10T12:58:02.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mon.b calling monitor election 2026-03-10T12:58:02.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mon.a calling monitor election 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mon.c calling monitor election 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: monmap epoch 3 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: min_mon_release 19 (squid) 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: election_strategy: 1 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: fsmap 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: mgrmap e12: a(active, since 27s) 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: overall HEALTH_OK 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: Deploying daemon mgr.b on vm06 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:02.848 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:02.860 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:02 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:02.771+0000 7fa3a72e5140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mon.b calling monitor election 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm04=a;vm06=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: Saving service mgr spec with placement vm04=a;vm06=b;count:2 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mon.b calling monitor election 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mon.a calling monitor election 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mon.c calling monitor election 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: monmap epoch 3 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: last_changed 2026-03-10T12:57:54.377522+0000 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: min_mon_release 19 (squid) 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: election_strategy: 1 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.c 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: 2: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: fsmap 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: mgrmap e12: a(active, since 27s) 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: overall HEALTH_OK 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: Deploying daemon mgr.b on vm06 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:02.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:02.893 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-10T12:58:02.893 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:02.893 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T12:58:02.893 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:02.893 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:02.894 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 12:57:43.004699660 +0000 2026-03-10T12:58:02.894 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 12:54:13.830429097 +0000 2026-03-10T12:58:02.894 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 12:54:13.830429097 +0000 2026-03-10T12:58:02.894 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 12:50:53.252000000 +0000 2026-03-10T12:58:02.894 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T12:58:02.962 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T12:58:02.962 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T12:58:02.962 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000152095 s, 3.4 MB/s 2026-03-10T12:58:02.963 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T12:58:03.025 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 12:57:43.033699663 +0000 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 12:54:13.846429099 +0000 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 12:54:13.846429099 +0000 2026-03-10T12:58:03.089 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 12:50:53.293000000 +0000 2026-03-10T12:58:03.089 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T12:58:03.156 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T12:58:03.156 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T12:58:03.156 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000188853 s, 2.7 MB/s 2026-03-10T12:58:03.157 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T12:58:03.234 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T12:58:03.234 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.103+0000 7fa3a72e5140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: from numpy import show_config as show_numpy_config 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.192+0000 7fa3a72e5140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-10T12:58:03.236 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.233+0000 7fa3a72e5140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-10T12:58:03.251 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T12:58:03.251 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-10T12:58:03.307 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-10T12:58:03.307 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-10T12:58:03.307 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-10T12:58:03.307 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-10T12:58:03.307 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-10T12:58:03.307 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T12:58:03.307 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T12:58:03.307 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 12:57:46.702499134 +0000 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 12:54:13.491952527 +0000 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 12:54:13.491952527 +0000 2026-03-10T12:58:03.368 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 12:50:16.226000000 +0000 2026-03-10T12:58:03.368 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T12:58:03.432 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T12:58:03.432 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T12:58:03.432 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000172132 s, 3.0 MB/s 2026-03-10T12:58:03.433 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T12:58:03.490 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-10T12:58:03.546 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-10T12:58:03.546 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:03.546 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 222 Links: 1 Device type: fc,20 2026-03-10T12:58:03.546 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:03.546 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:03.547 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 12:57:46.726499213 +0000 2026-03-10T12:58:03.547 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 12:54:13.447952481 +0000 2026-03-10T12:58:03.547 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 12:54:13.447952481 +0000 2026-03-10T12:58:03.547 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 12:50:16.230000000 +0000 2026-03-10T12:58:03.547 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T12:58:03.609 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.304+0000 7fa3a72e5140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-10T12:58:03.615 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T12:58:03.615 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T12:58:03.615 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000144431 s, 3.5 MB/s 2026-03-10T12:58:03.616 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T12:58:03.676 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,30 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:03.732 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 12:57:46.750499293 +0000 2026-03-10T12:58:03.733 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 12:54:13.491952527 +0000 2026-03-10T12:58:03.733 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 12:54:13.491952527 +0000 2026-03-10T12:58:03.733 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 12:50:16.238000000 +0000 2026-03-10T12:58:03.733 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T12:58:03.795 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T12:58:03.795 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T12:58:03.795 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000130284 s, 3.9 MB/s 2026-03-10T12:58:03.795 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T12:58:03.847 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:58:03.378+0000 7f2d5a854640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-10T12:58:03.856 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-10T12:58:03.913 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-10T12:58:03.913 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 248 Links: 1 Device type: fc,40 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 12:57:46.774499372 +0000 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 12:54:13.496952532 +0000 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 12:54:13.496952532 +0000 2026-03-10T12:58:03.914 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 12:50:16.254000000 +0000 2026-03-10T12:58:03.914 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T12:58:03.980 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T12:58:03.980 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T12:58:03.980 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000184766 s, 2.8 MB/s 2026-03-10T12:58:03.981 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T12:58:04.040 INFO:tasks.cephadm:Deploying osd.0 on vm04 with /dev/vde... 2026-03-10T12:58:04.040 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vde 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.805+0000 7fa3a72e5140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.912+0000 7fa3a72e5140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.951+0000 7fa3a72e5140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:03.985+0000 7fa3a72e5140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.025+0000 7fa3a72e5140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: Reconfiguring daemon mgr.a on vm04 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.065 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:04.066 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.066 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.066 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: Reconfiguring daemon mgr.a on vm04 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:04.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: Reconfiguring mgr.a (unknown last config time)... 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: Reconfiguring daemon mgr.a on vm04 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:04.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:04.206 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:58:04.359 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.063+0000 7fa3a72e5140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-10T12:58:04.359 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.231+0000 7fa3a72e5140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-10T12:58:04.359 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.283+0000 7fa3a72e5140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-10T12:58:04.780 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.504+0000 7fa3a72e5140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-10T12:58:05.022 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:58:05.041 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm04:/dev/vde 2026-03-10T12:58:05.054 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.778+0000 7fa3a72e5140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-10T12:58:05.054 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.815+0000 7fa3a72e5140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-10T12:58:05.054 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.856+0000 7fa3a72e5140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-10T12:58:05.054 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.933+0000 7fa3a72e5140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-10T12:58:05.054 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:04 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:04.971+0000 7fa3a72e5140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-10T12:58:05.211 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:58:05.318 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:05.052+0000 7fa3a72e5140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-10T12:58:05.318 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:05.159+0000 7fa3a72e5140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-10T12:58:05.318 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:05.294+0000 7fa3a72e5140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-10T12:58:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:05 vm09 ceph-mon[53016]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:05 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:05.610 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-b[56978]: 2026-03-10T12:58:05.330+0000 7fa3a72e5140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-10T12:58:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-mon[56046]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:05 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:05.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:05 vm04 ceph-mon[50620]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:05.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:05 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: Standby manager daemon b started 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]: dispatch 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]': finished 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:58:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:06.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: Standby manager daemon b started 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]': finished 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: Standby manager daemon b started 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.? 192.168.123.106:0/1312271984' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]: dispatch 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1278729686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "14c98e21-9eba-4d05-819e-01138b9e7efe"}]': finished 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: osdmap e5: 1 total, 0 up, 1 in 2026-03-10T12:58:06.610 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:07.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:07 vm09 ceph-mon[53016]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:07.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:07 vm09 ceph-mon[53016]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-10T12:58:07.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:07 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:58:07.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:07 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2214890176' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:07.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:07 vm06 ceph-mon[56046]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:07.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:07 vm06 ceph-mon[56046]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-10T12:58:07.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:07 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:58:07.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:07 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2214890176' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:07.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:07 vm04 ceph-mon[50620]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:07.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:07 vm04 ceph-mon[50620]: mgrmap e13: a(active, since 32s), standbys: b 2026-03-10T12:58:07.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:07 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-10T12:58:07.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:07 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2214890176' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:09.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:09 vm09 ceph-mon[53016]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:09.600 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:09 vm04 ceph-mon[50620]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:09.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:09 vm06 ceph-mon[56046]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:11.526 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:11 vm04 ceph-mon[50620]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:11.526 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:11 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:58:11.526 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:11 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:11.526 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:11 vm04 ceph-mon[50620]: Deploying daemon osd.0 on vm04 2026-03-10T12:58:11.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:11 vm09 ceph-mon[53016]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:11.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:11 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:58:11.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:11 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:11.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:11 vm09 ceph-mon[53016]: Deploying daemon osd.0 on vm04 2026-03-10T12:58:11.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:11 vm06 ceph-mon[56046]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:11.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:11 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:58:11.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:11 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:11.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:11 vm06 ceph-mon[56046]: Deploying daemon osd.0 on vm04 2026-03-10T12:58:13.582 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 0 on host 'vm04' 2026-03-10T12:58:13.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:13.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:13 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:13.622 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:13 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.649 DEBUG:teuthology.orchestra.run.vm04:osd.0> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.0.service 2026-03-10T12:58:13.651 INFO:tasks.cephadm:Deploying osd.1 on vm04 with /dev/vdd... 2026-03-10T12:58:13.651 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vdd 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:13 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:13.923 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:58:14.356 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:14 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:14.356 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:14 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.356 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:14 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:14 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:14.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:14 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:14 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:14 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:14 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:14 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:14.903 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:14 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T12:58:14.717+0000 7efd526f5740 -1 osd.0 0 log_to_monitors true 2026-03-10T12:58:15.277 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:58:15.295 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm04:/dev/vdd 2026-03-10T12:58:15.453 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: Detected new or changed devices on vm04 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: Unable to set osd_memory_target on vm04 to 269533593: error parsing value: Value '269533593' is below minimum 939524096 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:15.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:15 vm06 ceph-mon[56046]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:15.867 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: Detected new or changed devices on vm04 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: Unable to set osd_memory_target on vm04 to 269533593: error parsing value: Value '269533593' is below minimum 939524096 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:15.868 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:15 vm04 ceph-mon[50620]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: Detected new or changed devices on vm04 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm04 to 257.0M 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: Unable to set osd_memory_target on vm04 to 269533593: error parsing value: Value '269533593' is below minimum 939524096 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:16.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:15 vm09 ceph-mon[53016]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='client.24116 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:16.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/704975255' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]': finished 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:16.878 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:16 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='client.24116 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/704975255' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]': finished 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:17.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:16 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='client.24116 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/704975255' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a64d29f1-c452-42b0-9e75-cb16d2c94ef2"}]': finished 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:17.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:16 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:17 vm04 ceph-mon[50620]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:17.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:17 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2307819586' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:17.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:17 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:17.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:17 vm04 ceph-mon[50620]: from='osd.0 ' entity='osd.0' 2026-03-10T12:58:17.877 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T12:58:17.431+0000 7efd4e676640 -1 osd.0 0 waiting for initial osdmap 2026-03-10T12:58:17.877 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:17 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T12:58:17.436+0000 7efd49c9f640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:58:18.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:17 vm09 ceph-mon[53016]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:18.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:17 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2307819586' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:18.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:17 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:18.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:17 vm09 ceph-mon[53016]: from='osd.0 ' entity='osd.0' 2026-03-10T12:58:18.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:17 vm06 ceph-mon[56046]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:18.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:17 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2307819586' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:18.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:17 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:18.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:17 vm06 ceph-mon[56046]: from='osd.0 ' entity='osd.0' 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144] boot 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:18 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144] boot 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:18 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:19.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:58:19.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:58:19.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: osd.0 [v2:192.168.123.104:6802/276222144,v1:192.168.123.104:6803/276222144] boot 2026-03-10T12:58:19.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: osdmap e8: 2 total, 1 up, 2 in 2026-03-10T12:58:19.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T12:58:19.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:18 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:19.948 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:19 vm04 ceph-mon[50620]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:20.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:19 vm09 ceph-mon[53016]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:20.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:19 vm06 ceph-mon[56046]: pgmap v19: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T12:58:20.787 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:20 vm04 ceph-mon[50620]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T12:58:20.787 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:20.787 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:58:20.787 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:21.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:20 vm09 ceph-mon[53016]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T12:58:21.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:20 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:21.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:20 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:58:21.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:20 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:21.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:20 vm06 ceph-mon[56046]: osdmap e9: 2 total, 1 up, 2 in 2026-03-10T12:58:21.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:21.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:58:21.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:22.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:21 vm09 ceph-mon[53016]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:22.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:21 vm09 ceph-mon[53016]: Deploying daemon osd.1 on vm04 2026-03-10T12:58:22.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:21 vm06 ceph-mon[56046]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:22.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:21 vm06 ceph-mon[56046]: Deploying daemon osd.1 on vm04 2026-03-10T12:58:22.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:21 vm04 ceph-mon[50620]: pgmap v22: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:22.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:21 vm04 ceph-mon[50620]: Deploying daemon osd.1 on vm04 2026-03-10T12:58:22.834 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:22.834 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:22.834 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:23.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:22 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:23.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:22 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:23.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:22 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:23.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:23.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:23.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:23.371 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 1 on host 'vm04' 2026-03-10T12:58:23.431 DEBUG:teuthology.orchestra.run.vm04:osd.1> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.1.service 2026-03-10T12:58:23.433 INFO:tasks.cephadm:Deploying osd.2 on vm06 with /dev/vde... 2026-03-10T12:58:23.433 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vde 2026-03-10T12:58:23.591 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.b/config 2026-03-10T12:58:23.983 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T12:58:23.980+0000 7f6110e7d740 -1 osd.1 0 log_to_monitors true 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.234 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:24 vm04 ceph-mon[50620]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.290 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:24 vm06 ceph-mon[56046]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: pgmap v23: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:24.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:24 vm09 ceph-mon[53016]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T12:58:24.415 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:58:24.431 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm06:/dev/vde 2026-03-10T12:58:24.600 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.b/config 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: Detected new or changed devices on vm04 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm04 to 128.5M 2026-03-10T12:58:25.300 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: Unable to set osd_memory_target on vm04 to 134766796: error parsing value: Value '134766796' is below minimum 939524096 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: Detected new or changed devices on vm04 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm04 to 128.5M 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: Unable to set osd_memory_target on vm04 to 134766796: error parsing value: Value '134766796' is below minimum 939524096 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:25.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:25 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: Detected new or changed devices on vm04 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: pgmap v24: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm04 to 128.5M 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: Unable to set osd_memory_target on vm04 to 134766796: error parsing value: Value '134766796' is below minimum 939524096 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: osdmap e10: 2 total, 1 up, 2 in 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='client.24145 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:25.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T12:58:25.387+0000 7f610d611640 -1 osd.1 0 waiting for initial osdmap 2026-03-10T12:58:25.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:25 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T12:58:25.391+0000 7f6108c28640 -1 osd.1 11 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]: dispatch 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]': finished 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865] boot 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:26.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:26 vm09 ceph-mon[53016]: from='client.? 192.168.123.106:0/1263457982' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]': finished 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865] boot 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:26.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:26 vm06 ceph-mon[56046]: from='client.? 192.168.123.106:0/1263457982' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:26.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-10T12:58:26.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: osdmap e11: 2 total, 1 up, 2 in 2026-03-10T12:58:26.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]: dispatch 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='client.? 192.168.123.106:0/1809959544' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "54987266-93f3-44da-bf3e-94dfc1757581"}]': finished 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: osd.1 [v2:192.168.123.104:6810/2339889865,v1:192.168.123.104:6811/2339889865] boot 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: osdmap e12: 3 total, 2 up, 3 in 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:26.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:26 vm04 ceph-mon[50620]: from='client.? 192.168.123.106:0/1263457982' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:27.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:27 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:58:27.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:27 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:58:27.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:27 vm09 ceph-mon[53016]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:27.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:27 vm09 ceph-mon[53016]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T12:58:27.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:27 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:27.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:27 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:58:27.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:27 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:58:27.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:27 vm06 ceph-mon[56046]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:27.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:27 vm06 ceph-mon[56046]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T12:58:27.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:27 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:27.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:27 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:58:27.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:27 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:58:27.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:27 vm04 ceph-mon[50620]: pgmap v28: 0 pgs: ; 0 B data, 426 MiB used, 20 GiB / 20 GiB avail 2026-03-10T12:58:27.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:27 vm04 ceph-mon[50620]: osdmap e13: 3 total, 2 up, 3 in 2026-03-10T12:58:27.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:27 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:29.783 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:29 vm06 ceph-mon[56046]: pgmap v30: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:29.783 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T12:58:29.783 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:29.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:29 vm04 ceph-mon[50620]: pgmap v30: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:29.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T12:58:29.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:30.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:29 vm09 ceph-mon[53016]: pgmap v30: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:30.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:29 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T12:58:30.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:29 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:30.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:30 vm06 ceph-mon[56046]: Deploying daemon osd.2 on vm06 2026-03-10T12:58:30.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:30 vm04 ceph-mon[50620]: Deploying daemon osd.2 on vm06 2026-03-10T12:58:31.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:30 vm09 ceph-mon[53016]: Deploying daemon osd.2 on vm06 2026-03-10T12:58:31.636 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:31 vm06 ceph-mon[56046]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:31.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:31 vm04 ceph-mon[50620]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:32.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:31 vm09 ceph-mon[53016]: pgmap v31: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:32.677 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 2 on host 'vm06' 2026-03-10T12:58:32.734 DEBUG:teuthology.orchestra.run.vm06:osd.2> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.2.service 2026-03-10T12:58:32.735 INFO:tasks.cephadm:Deploying osd.3 on vm06 with /dev/vdd... 2026-03-10T12:58:32.735 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vdd 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:32.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:32 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:32.929 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:32 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.018 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.b/config 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:33.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:32 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.610 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:33 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T12:58:33.325+0000 7f696c065740 -1 osd.2 0 log_to_monitors true 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:33.868 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:33 vm06 ceph-mon[56046]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:33.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:33.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:33.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:33.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:33.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:33 vm04 ceph-mon[50620]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:34.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:33 vm09 ceph-mon[53016]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T12:58:34.400 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T12:58:34.419 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm06:/dev/vdd 2026-03-10T12:58:34.592 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.b/config 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: Detected new or changed devices on vm06 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:35.046 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:34 vm06 ceph-mon[56046]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: Detected new or changed devices on vm06 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:35.083 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:34 vm04 ceph-mon[50620]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: Detected new or changed devices on vm06 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm06 to 257.0M 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: Unable to set osd_memory_target on vm06 to 269536460: error parsing value: Value '269536460' is below minimum 939524096 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: osdmap e14: 3 total, 2 up, 3 in 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:35.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:34 vm09 ceph-mon[53016]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='client.24167 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='client.? 192.168.123.106:0/625014213' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]': finished 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: osdmap e16: 4 total, 2 up, 4 in 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.051 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:35 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='client.24167 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='client.? 192.168.123.106:0/625014213' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]': finished 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: osdmap e16: 4 total, 2 up, 4 in 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:35 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='client.24167 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm06:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: osdmap e15: 3 total, 2 up, 3 in 2026-03-10T12:58:36.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='client.? 192.168.123.106:0/625014213' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]: dispatch 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e5eb8885-521e-4b6c-8370-7d8604264779"}]': finished 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: osdmap e16: 4 total, 2 up, 4 in 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:36.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:35 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:36.359 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:36 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T12:58:36.219+0000 7f6967fe6640 -1 osd.2 0 waiting for initial osdmap 2026-03-10T12:58:36.360 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:36 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T12:58:36.225+0000 7f696360f640 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:58:37.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:36 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:58:37.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:36 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:58:37.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:36 vm09 ceph-mon[53016]: from='client.? 192.168.123.106:0/687468103' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:37.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:36 vm09 ceph-mon[53016]: from='osd.2 ' entity='osd.2' 2026-03-10T12:58:37.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:36 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:37.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:36 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:58:37.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:36 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:58:37.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:36 vm06 ceph-mon[56046]: from='client.? 192.168.123.106:0/687468103' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:37.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:36 vm06 ceph-mon[56046]: from='osd.2 ' entity='osd.2' 2026-03-10T12:58:37.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:37.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:36 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:58:37.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:36 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:58:37.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:36 vm04 ceph-mon[50620]: from='client.? 192.168.123.106:0/687468103' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:37.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:36 vm04 ceph-mon[50620]: from='osd.2 ' entity='osd.2' 2026-03-10T12:58:37.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716] boot 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:37 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:38.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:38.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716] boot 2026-03-10T12:58:38.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T12:58:38.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:38.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T12:58:38.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: osd.2 [v2:192.168.123.106:6800/285173716,v1:192.168.123.106:6801/285173716] boot 2026-03-10T12:58:38.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: osdmap e17: 4 total, 3 up, 4 in 2026-03-10T12:58:38.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T12:58:38.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:39.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:39 vm06 ceph-mon[56046]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T12:58:39.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:39.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:39 vm06 ceph-mon[56046]: pgmap v40: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:39.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:39.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:39 vm09 ceph-mon[53016]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T12:58:39.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:39 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:39.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:39 vm09 ceph-mon[53016]: pgmap v40: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:39.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:39 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:39.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:39 vm04 ceph-mon[50620]: osdmap e18: 4 total, 3 up, 4 in 2026-03-10T12:58:39.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:39.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:39 vm04 ceph-mon[50620]: pgmap v40: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:39.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:40.406 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 ceph-mon[56046]: Deploying daemon osd.3 on vm06 2026-03-10T12:58:40.406 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64608]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-10T12:58:40.406 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64608]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.406 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64608]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.406 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64608]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 ceph-mon[53016]: Deploying daemon osd.3 on vm06 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 sudo[54446]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 sudo[54446]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 sudo[54446]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:40 vm09 sudo[54446]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:40.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71047]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-10T12:58:40.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71047]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71047]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71047]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:40.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71051]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdd 2026-03-10T12:58:40.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71051]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71051]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.627 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71051]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: osdmap e19: 4 total, 3 up, 4 in 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 ceph-mon[50620]: Deploying daemon osd.3 on vm06 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71055]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71055]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71055]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:40 vm04 sudo[71055]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:40.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64634]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T12:58:40.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64634]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T12:58:40.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64634]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T12:58:40.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:40 vm06 sudo[64634]: pam_unix(sudo:session): session closed for user root 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: pgmap v43: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:41 vm09 ceph-mon[53016]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: pgmap v43: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:41 vm06 ceph-mon[56046]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: osdmap e20: 4 total, 3 up, 4 in 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: pgmap v43: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T12:58:41.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:41 vm04 ceph-mon[50620]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:42 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: osdmap e21: 4 total, 3 up, 4 in 2026-03-10T12:58:42.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:42.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:42.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:42.632 INFO:teuthology.orchestra.run.vm06.stdout:Created osd(s) 3 on host 'vm06' 2026-03-10T12:58:42.683 DEBUG:teuthology.orchestra.run.vm06:osd.3> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.3.service 2026-03-10T12:58:42.684 INFO:tasks.cephadm:Deploying osd.4 on vm09 with /dev/vde... 2026-03-10T12:58:42.684 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vde 2026-03-10T12:58:42.852 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.498 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:43 vm06 ceph-mon[56046]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.498 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 12:58:43 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T12:58:43.170+0000 7ffb1778f740 -1 osd.3 0 log_to_monitors true 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.560 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:43 vm09 ceph-mon[53016]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: pgmap v45: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:43.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:43.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:43 vm04 ceph-mon[50620]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T12:58:43.656 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:58:43.675 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm09:/dev/vde 2026-03-10T12:58:43.835 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: mgrmap e14: a(active, since 69s), standbys: b 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: Detected new or changed devices on vm06 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm06 to 128.5M 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: Unable to set osd_memory_target on vm06 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='client.24193 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:44.598 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:44 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: mgrmap e14: a(active, since 69s), standbys: b 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: Detected new or changed devices on vm06 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm06 to 128.5M 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: Unable to set osd_memory_target on vm06 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='client.24193 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:44 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:44.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: mgrmap e14: a(active, since 69s), standbys: b 2026-03-10T12:58:44.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: Detected new or changed devices on vm06 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm06 to 128.5M 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: Unable to set osd_memory_target on vm06 to 134768230: error parsing value: Value '134768230' is below minimum 939524096 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: osdmap e22: 4 total, 3 up, 4 in 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='client.24193 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:44.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:44 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='client.? 192.168.123.109:0/96238429' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]': finished 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: osdmap e24: 5 total, 3 up, 5 in 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='client.? 192.168.123.109:0/2144219244' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:46.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:45 vm09 ceph-mon[53016]: from='osd.3 ' entity='osd.3' 2026-03-10T12:58:46.110 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 12:58:45 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T12:58:45.615+0000 7ffb13f23640 -1 osd.3 0 waiting for initial osdmap 2026-03-10T12:58:46.110 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 12:58:45 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T12:58:45.622+0000 7ffb0ed39640 -1 osd.3 24 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='client.? 192.168.123.109:0/96238429' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]': finished 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: osdmap e24: 5 total, 3 up, 5 in 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='client.? 192.168.123.109:0/2144219244' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:46.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:45 vm06 ceph-mon[56046]: from='osd.3 ' entity='osd.3' 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: pgmap v47: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: osdmap e23: 4 total, 3 up, 4 in 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/96238429' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]: dispatch 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "1c675835-f3b7-40fb-a62f-e795b68d3dcb"}]': finished 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: osdmap e24: 5 total, 3 up, 5 in 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/2144219244' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:46.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:45 vm04 ceph-mon[50620]: from='osd.3 ' entity='osd.3' 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640] boot 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:46 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640] boot 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:46 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: osd.3 [v2:192.168.123.106:6808/2984060640,v1:192.168.123.106:6809/2984060640] boot 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: osdmap e25: 5 total, 4 up, 5 in 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T12:58:47.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:46 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:48.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:47 vm09 ceph-mon[53016]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:48.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:47 vm09 ceph-mon[53016]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T12:58:48.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:47 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:48.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:47 vm06 ceph-mon[56046]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:48.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:47 vm06 ceph-mon[56046]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T12:58:48.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:47 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:48.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:47 vm04 ceph-mon[50620]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-10T12:58:48.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:47 vm04 ceph-mon[50620]: osdmap e26: 5 total, 4 up, 5 in 2026-03-10T12:58:48.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:47 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:49.065 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:48 vm09 ceph-mon[53016]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T12:58:49.065 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:48 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:49.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:48 vm06 ceph-mon[56046]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T12:58:49.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:48 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:49.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:48 vm04 ceph-mon[50620]: osdmap e27: 5 total, 4 up, 5 in 2026-03-10T12:58:49.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:48 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:49.898 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:49 vm09 ceph-mon[53016]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:49.898 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:49 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T12:58:49.899 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:49 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:49.899 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:49 vm09 ceph-mon[53016]: Deploying daemon osd.4 on vm09 2026-03-10T12:58:50.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:49 vm06 ceph-mon[56046]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:50.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:49 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T12:58:50.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:49 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:50.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:49 vm06 ceph-mon[56046]: Deploying daemon osd.4 on vm09 2026-03-10T12:58:50.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:49 vm04 ceph-mon[50620]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:50.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:49 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T12:58:50.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:49 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:50.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:49 vm04 ceph-mon[50620]: Deploying daemon osd.4 on vm09 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:51 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:58:52.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:52.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.241 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 4 on host 'vm09' 2026-03-10T12:58:52.298 DEBUG:teuthology.orchestra.run.vm09:osd.4> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.4.service 2026-03-10T12:58:52.300 INFO:tasks.cephadm:Deploying osd.5 on vm09 with /dev/vdd... 2026-03-10T12:58:52.300 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- lvm zap /dev/vdd 2026-03-10T12:58:52.477 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:58:52 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:58:52.473+0000 7f6277bf3740 -1 osd.4 0 log_to_monitors true 2026-03-10T12:58:52.587 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:52.818 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:52 vm09 ceph-mon[53016]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:53.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:52 vm06 ceph-mon[56046]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:53.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:53.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:52 vm04 ceph-mon[50620]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T12:58:53.845 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T12:58:53.862 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph orch daemon add osd vm09:/dev/vdd 2026-03-10T12:58:54.019 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.c/config 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.041 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:53 vm09 ceph-mon[53016]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.109 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:53 vm06 ceph-mon[56046]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: osdmap e28: 5 total, 4 up, 5 in 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:53 vm04 ceph-mon[50620]: from='osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: Detected new or changed devices on vm09 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm09 to 4353M 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: osdmap e29: 5 total, 4 up, 5 in 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:54.833 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:54 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: Detected new or changed devices on vm09 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm09 to 4353M 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: osdmap e29: 5 total, 4 up, 5 in 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:55.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:54 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: Detected new or changed devices on vm09 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm09 to 4353M 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: osdmap e29: 5 total, 4 up, 5 in 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:58:55.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:54 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.549 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:58:55 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:58:55.303+0000 7f6274387640 -1 osd.4 0 waiting for initial osdmap 2026-03-10T12:58:55.549 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:58:55 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:58:55.311+0000 7f626f19d640 -1 osd.4 30 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='client.24217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='client.? 192.168.123.109:0/2914335265' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]': finished 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: osdmap e30: 6 total, 4 up, 6 in 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:55.828 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:56.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:56.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='osd.4 ' entity='osd.4' 2026-03-10T12:58:56.097 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:55 vm09 ceph-mon[53016]: from='client.? 192.168.123.109:0/718000084' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='client.24217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='client.? 192.168.123.109:0/2914335265' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]': finished 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: osdmap e30: 6 total, 4 up, 6 in 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='osd.4 ' entity='osd.4' 2026-03-10T12:58:56.110 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:55 vm06 ceph-mon[56046]: from='client.? 192.168.123.109:0/718000084' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:56.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:58:56.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:58:56.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='client.24217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm09:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:58:56.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/2914335265' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]: dispatch 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "329c35de-29d1-40e7-bb6f-b4581bee34b7"}]': finished 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: osdmap e30: 6 total, 4 up, 6 in 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='osd.4 ' entity='osd.4' 2026-03-10T12:58:56.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:55 vm04 ceph-mon[50620]: from='client.? 192.168.123.109:0/718000084' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T12:58:57.359 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:57 vm06 ceph-mon[56046]: osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420] boot 2026-03-10T12:58:57.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:57 vm06 ceph-mon[56046]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T12:58:57.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:57 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:57.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:57 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:57.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:57 vm04 ceph-mon[50620]: osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420] boot 2026-03-10T12:58:57.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:57 vm04 ceph-mon[50620]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T12:58:57.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:57 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:57.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:57 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:57.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:57 vm09 ceph-mon[53016]: osd.4 [v2:192.168.123.109:6800/726545420,v1:192.168.123.109:6801/726545420] boot 2026-03-10T12:58:57.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:57 vm09 ceph-mon[53016]: osdmap e31: 6 total, 5 up, 6 in 2026-03-10T12:58:57.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:57 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T12:58:57.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:57 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:58.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:58 vm04 ceph-mon[50620]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:58:58.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:58 vm04 ceph-mon[50620]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T12:58:58.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:58 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:58.467 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:58 vm09 ceph-mon[53016]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:58:58.467 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:58 vm09 ceph-mon[53016]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T12:58:58.467 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:58 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:58.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:58 vm06 ceph-mon[56046]: pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:58:58.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:58 vm06 ceph-mon[56046]: osdmap e32: 6 total, 5 up, 6 in 2026-03-10T12:58:58.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:58 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:59.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:59 vm09 ceph-mon[53016]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T12:58:59.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:58:59 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:59.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:59 vm06 ceph-mon[56046]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T12:58:59.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:58:59 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:58:59.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:59 vm04 ceph-mon[50620]: osdmap e33: 6 total, 5 up, 6 in 2026-03-10T12:58:59.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:58:59 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:00.157 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:00 vm09 ceph-mon[53016]: pgmap v65: 1 pgs: 1 remapped+peering; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:00.157 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:00.157 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:00 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:00.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:00 vm06 ceph-mon[56046]: pgmap v65: 1 pgs: 1 remapped+peering; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:00.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:00 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:00.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:00 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:00.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:00 vm04 ceph-mon[50620]: pgmap v65: 1 pgs: 1 remapped+peering; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:00.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:00.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:00 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:01.243 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:01 vm09 ceph-mon[53016]: Deploying daemon osd.5 on vm09 2026-03-10T12:59:01.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:01 vm06 ceph-mon[56046]: Deploying daemon osd.5 on vm09 2026-03-10T12:59:01.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:01 vm04 ceph-mon[50620]: Deploying daemon osd.5 on vm09 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: pgmap v66: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:02.347 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:02 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.502 INFO:teuthology.orchestra.run.vm09.stdout:Created osd(s) 5 on host 'vm09' 2026-03-10T12:59:02.577 DEBUG:teuthology.orchestra.run.vm09:osd.5> sudo journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.5.service 2026-03-10T12:59:02.578 INFO:tasks.cephadm:Waiting for 6 OSDs to come up... 2026-03-10T12:59:02.578 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd stat -f json 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: pgmap v66: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:02.601 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:02 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: pgmap v66: 1 pgs: 1 remapped+peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:02 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:02.755 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:02.991 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:03.037 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":33,"num_osds":6,"num_up_osds":5,"osd_up_since":1773147536,"num_in_osds":6,"osd_in_since":1773147535,"num_remapped_pgs":0} 2026-03-10T12:59:03.097 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:02 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:02.839+0000 7f255cd40740 -1 osd.5 0 log_to_monitors true 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.400 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:03 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2684364856' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:03 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2684364856' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:03.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:03.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:03.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T12:59:03.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:03 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2684364856' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:04.037 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd stat -f json 2026-03-10T12:59:04.200 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 72 KiB/s, 0 objects/s recovering 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: osdmap e34: 6 total, 5 up, 6 in 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:04.318 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:04 vm04 ceph-mon[50620]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:04.424 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:04.494 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":34,"num_osds":6,"num_up_osds":5,"osd_up_since":1773147536,"num_in_osds":6,"osd_in_since":1773147535,"num_remapped_pgs":0} 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 72 KiB/s, 0 objects/s recovering 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: osdmap e34: 6 total, 5 up, 6 in 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:04.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:04 vm09 ceph-mon[53016]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 72 KiB/s, 0 objects/s recovering 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-10T12:59:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: osdmap e34: 6 total, 5 up, 6 in 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:04.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:04 vm06 ceph-mon[56046]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T12:59:05.495 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd stat -f json 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: Detected new or changed devices on vm09 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: Adjusting osd_memory_target on vm09 to 2176M 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1793514351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: osdmap e35: 6 total, 5 up, 6 in 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.524 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:05 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: Detected new or changed devices on vm09 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: Adjusting osd_memory_target on vm09 to 2176M 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1793514351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: osdmap e35: 6 total, 5 up, 6 in 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:05 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: Detected new or changed devices on vm09 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: Adjusting osd_memory_target on vm09 to 2176M 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1793514351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: osdmap e35: 6 total, 5 up, 6 in 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:05 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:05.671 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:05.883 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:05.930 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":36,"num_osds":6,"num_up_osds":5,"osd_up_since":1773147536,"num_in_osds":6,"osd_in_since":1773147535,"num_remapped_pgs":1} 2026-03-10T12:59:06.098 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:05 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:05.599+0000 7f25594d4640 -1 osd.5 0 waiting for initial osdmap 2026-03-10T12:59:06.098 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:05 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:05.609+0000 7f25542ea640 -1 osd.5 36 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: purged_snaps scrub starts 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: purged_snaps scrub ok 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: osdmap e36: 6 total, 5 up, 6 in 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: from='osd.5 ' entity='osd.5' 2026-03-10T12:59:06.597 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:06 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2822795351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: purged_snaps scrub starts 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: osdmap e36: 6 total, 5 up, 6 in 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: from='osd.5 ' entity='osd.5' 2026-03-10T12:59:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:06 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2822795351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:06.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: purged_snaps scrub starts 2026-03-10T12:59:06.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 62 KiB/s, 0 objects/s recovering 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: osdmap e36: 6 total, 5 up, 6 in 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: from='osd.5 ' entity='osd.5' 2026-03-10T12:59:06.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:06 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2822795351' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:06.931 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd stat -f json 2026-03-10T12:59:07.096 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:07.302 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:07.349 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":37,"num_osds":6,"num_up_osds":6,"osd_up_since":1773147546,"num_in_osds":6,"osd_in_since":1773147535,"num_remapped_pgs":1} 2026-03-10T12:59:07.349 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd dump --format=json 2026-03-10T12:59:07.506 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:07.520 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:07 vm04 ceph-mon[50620]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:07.736 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:07.736 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":38,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","created":"2026-03-10T12:57:12.584951+0000","modified":"2026-03-10T12:59:07.507943+0000","last_up_change":"2026-03-10T12:59:06.504723+0000","last_in_change":"2026-03-10T12:58:55.082674+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T12:58:38.326337+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"14c98e21-9eba-4d05-819e-01138b9e7efe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6803","nonce":276222144}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6805","nonce":276222144}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6809","nonce":276222144}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6807","nonce":276222144}]},"public_addr":"192.168.123.104:6803/276222144","cluster_addr":"192.168.123.104:6805/276222144","heartbeat_back_addr":"192.168.123.104:6809/276222144","heartbeat_front_addr":"192.168.123.104:6807/276222144","state":["exists","up"]},{"osd":1,"uuid":"a64d29f1-c452-42b0-9e75-cb16d2c94ef2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6811","nonce":2339889865}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6813","nonce":2339889865}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6817","nonce":2339889865}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6815","nonce":2339889865}]},"public_addr":"192.168.123.104:6811/2339889865","cluster_addr":"192.168.123.104:6813/2339889865","heartbeat_back_addr":"192.168.123.104:6817/2339889865","heartbeat_front_addr":"192.168.123.104:6815/2339889865","state":["exists","up"]},{"osd":2,"uuid":"54987266-93f3-44da-bf3e-94dfc1757581","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6801","nonce":285173716}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6803","nonce":285173716}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6807","nonce":285173716}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6805","nonce":285173716}]},"public_addr":"192.168.123.106:6801/285173716","cluster_addr":"192.168.123.106:6803/285173716","heartbeat_back_addr":"192.168.123.106:6807/285173716","heartbeat_front_addr":"192.168.123.106:6805/285173716","state":["exists","up"]},{"osd":3,"uuid":"e5eb8885-521e-4b6c-8370-7d8604264779","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":37,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6809","nonce":2984060640}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6811","nonce":2984060640}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6815","nonce":2984060640}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6813","nonce":2984060640}]},"public_addr":"192.168.123.106:6809/2984060640","cluster_addr":"192.168.123.106:6811/2984060640","heartbeat_back_addr":"192.168.123.106:6815/2984060640","heartbeat_front_addr":"192.168.123.106:6813/2984060640","state":["exists","up"]},{"osd":4,"uuid":"1c675835-f3b7-40fb-a62f-e795b68d3dcb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6801","nonce":726545420}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6803","nonce":726545420}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6807","nonce":726545420}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6805","nonce":726545420}]},"public_addr":"192.168.123.109:6801/726545420","cluster_addr":"192.168.123.109:6803/726545420","heartbeat_back_addr":"192.168.123.109:6807/726545420","heartbeat_front_addr":"192.168.123.109:6805/726545420","state":["exists","up"]},{"osd":5,"uuid":"329c35de-29d1-40e7-bb6f-b4581bee34b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6809","nonce":3212082069}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6811","nonce":3212082069}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6815","nonce":3212082069}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6813","nonce":3212082069}]},"public_addr":"192.168.123.109:6809/3212082069","cluster_addr":"192.168.123.109:6811/3212082069","heartbeat_back_addr":"192.168.123.109:6815/3212082069","heartbeat_front_addr":"192.168.123.109:6813/3212082069","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:15.734917+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:24.957450+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:34.284347+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:44.190420+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:53.436462+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:59:03.806044+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.104:6801/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/2118987935":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6801/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/3807242158":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2934451579":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/1685098299":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2928968844":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/3806698536":"2026-03-11T12:57:23.974125+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:59:07.784 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:07 vm04 ceph-mon[50620]: osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069] boot 2026-03-10T12:59:07.784 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:07 vm04 ceph-mon[50620]: osdmap e37: 6 total, 6 up, 6 in 2026-03-10T12:59:07.784 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:07 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:07.784 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:07 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1158670052' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:07.788 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T12:58:38.326337+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '21', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 6, 'score_stable': 6, 'optimal_score': 0.5, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T12:59:07.788 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd pool get .mgr pg_num 2026-03-10T12:59:07.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:07 vm09 ceph-mon[53016]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:07.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:07 vm09 ceph-mon[53016]: osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069] boot 2026-03-10T12:59:07.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:07 vm09 ceph-mon[53016]: osdmap e37: 6 total, 6 up, 6 in 2026-03-10T12:59:07.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:07 vm09 ceph-mon[53016]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:07.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:07 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1158670052' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:07.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:07 vm06 ceph-mon[56046]: pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:07.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:07 vm06 ceph-mon[56046]: osd.5 [v2:192.168.123.109:6808/3212082069,v1:192.168.123.109:6809/3212082069] boot 2026-03-10T12:59:07.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:07 vm06 ceph-mon[56046]: osdmap e37: 6 total, 6 up, 6 in 2026-03-10T12:59:07.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:07 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T12:59:07.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:07 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1158670052' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T12:59:07.948 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:08.163 INFO:teuthology.orchestra.run.vm04.stdout:pg_num: 1 2026-03-10T12:59:08.232 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T12:59:08.232 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T12:59:08.232 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T12:59:08.232 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph mgr dump --format=json 2026-03-10T12:59:08.396 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:08.626 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:08.669 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:08 vm04 ceph-mon[50620]: osdmap e38: 6 total, 6 up, 6 in 2026-03-10T12:59:08.669 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:08 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/4148603437' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:08.669 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:08 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1177107592' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T12:59:08.691 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":14,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":1086729154},{"type":"v1","addr":"192.168.123.104:6801","nonce":1086729154}]},"active_addr":"192.168.123.104:6801/1086729154","active_change":"2026-03-10T12:57:34.270833+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14211,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.104:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.104:0","nonce":290786436}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.104:0","nonce":1891794843}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.104:0","nonce":3333592019}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.104:0","nonce":2354347702}]}]} 2026-03-10T12:59:08.692 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T12:59:08.692 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T12:59:08.692 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd dump --format=json 2026-03-10T12:59:08.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:08 vm09 ceph-mon[53016]: osdmap e38: 6 total, 6 up, 6 in 2026-03-10T12:59:08.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:08 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/4148603437' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:08.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:08 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1177107592' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T12:59:08.854 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:08.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:08 vm06 ceph-mon[56046]: osdmap e38: 6 total, 6 up, 6 in 2026-03-10T12:59:08.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:08 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/4148603437' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:08.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:08 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1177107592' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T12:59:09.058 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:09.058 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":39,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","created":"2026-03-10T12:57:12.584951+0000","modified":"2026-03-10T12:59:08.519066+0000","last_up_change":"2026-03-10T12:59:06.504723+0000","last_in_change":"2026-03-10T12:58:55.082674+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T12:58:38.326337+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"14c98e21-9eba-4d05-819e-01138b9e7efe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6803","nonce":276222144}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6805","nonce":276222144}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6809","nonce":276222144}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6807","nonce":276222144}]},"public_addr":"192.168.123.104:6803/276222144","cluster_addr":"192.168.123.104:6805/276222144","heartbeat_back_addr":"192.168.123.104:6809/276222144","heartbeat_front_addr":"192.168.123.104:6807/276222144","state":["exists","up"]},{"osd":1,"uuid":"a64d29f1-c452-42b0-9e75-cb16d2c94ef2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6811","nonce":2339889865}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6813","nonce":2339889865}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6817","nonce":2339889865}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6815","nonce":2339889865}]},"public_addr":"192.168.123.104:6811/2339889865","cluster_addr":"192.168.123.104:6813/2339889865","heartbeat_back_addr":"192.168.123.104:6817/2339889865","heartbeat_front_addr":"192.168.123.104:6815/2339889865","state":["exists","up"]},{"osd":2,"uuid":"54987266-93f3-44da-bf3e-94dfc1757581","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6801","nonce":285173716}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6803","nonce":285173716}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6807","nonce":285173716}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6805","nonce":285173716}]},"public_addr":"192.168.123.106:6801/285173716","cluster_addr":"192.168.123.106:6803/285173716","heartbeat_back_addr":"192.168.123.106:6807/285173716","heartbeat_front_addr":"192.168.123.106:6805/285173716","state":["exists","up"]},{"osd":3,"uuid":"e5eb8885-521e-4b6c-8370-7d8604264779","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6809","nonce":2984060640}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6811","nonce":2984060640}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6815","nonce":2984060640}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6813","nonce":2984060640}]},"public_addr":"192.168.123.106:6809/2984060640","cluster_addr":"192.168.123.106:6811/2984060640","heartbeat_back_addr":"192.168.123.106:6815/2984060640","heartbeat_front_addr":"192.168.123.106:6813/2984060640","state":["exists","up"]},{"osd":4,"uuid":"1c675835-f3b7-40fb-a62f-e795b68d3dcb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6801","nonce":726545420}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6803","nonce":726545420}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6807","nonce":726545420}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6805","nonce":726545420}]},"public_addr":"192.168.123.109:6801/726545420","cluster_addr":"192.168.123.109:6803/726545420","heartbeat_back_addr":"192.168.123.109:6807/726545420","heartbeat_front_addr":"192.168.123.109:6805/726545420","state":["exists","up"]},{"osd":5,"uuid":"329c35de-29d1-40e7-bb6f-b4581bee34b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6809","nonce":3212082069}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6811","nonce":3212082069}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6815","nonce":3212082069}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6813","nonce":3212082069}]},"public_addr":"192.168.123.109:6809/3212082069","cluster_addr":"192.168.123.109:6811/3212082069","heartbeat_back_addr":"192.168.123.109:6815/3212082069","heartbeat_front_addr":"192.168.123.109:6813/3212082069","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:15.734917+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:24.957450+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:34.284347+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:44.190420+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:53.436462+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:59:03.806044+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.104:6801/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/2118987935":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6801/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/3807242158":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2934451579":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/1685098299":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2928968844":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/3806698536":"2026-03-11T12:57:23.974125+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:59:09.117 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T12:59:09.117 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd dump --format=json 2026-03-10T12:59:09.275 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:09.492 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:09.492 INFO:teuthology.orchestra.run.vm04.stdout:{"epoch":39,"fsid":"74d0e8a6-1c80-11f1-9542-4d7926099318","created":"2026-03-10T12:57:12.584951+0000","modified":"2026-03-10T12:59:08.519066+0000","last_up_change":"2026-03-10T12:59:06.504723+0000","last_in_change":"2026-03-10T12:58:55.082674+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T12:58:38.326337+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"14c98e21-9eba-4d05-819e-01138b9e7efe","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6803","nonce":276222144}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6805","nonce":276222144}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6809","nonce":276222144}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":276222144},{"type":"v1","addr":"192.168.123.104:6807","nonce":276222144}]},"public_addr":"192.168.123.104:6803/276222144","cluster_addr":"192.168.123.104:6805/276222144","heartbeat_back_addr":"192.168.123.104:6809/276222144","heartbeat_front_addr":"192.168.123.104:6807/276222144","state":["exists","up"]},{"osd":1,"uuid":"a64d29f1-c452-42b0-9e75-cb16d2c94ef2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":12,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6811","nonce":2339889865}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6813","nonce":2339889865}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6817","nonce":2339889865}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":2339889865},{"type":"v1","addr":"192.168.123.104:6815","nonce":2339889865}]},"public_addr":"192.168.123.104:6811/2339889865","cluster_addr":"192.168.123.104:6813/2339889865","heartbeat_back_addr":"192.168.123.104:6817/2339889865","heartbeat_front_addr":"192.168.123.104:6815/2339889865","state":["exists","up"]},{"osd":2,"uuid":"54987266-93f3-44da-bf3e-94dfc1757581","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":19,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6801","nonce":285173716}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6803","nonce":285173716}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6807","nonce":285173716}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":285173716},{"type":"v1","addr":"192.168.123.106:6805","nonce":285173716}]},"public_addr":"192.168.123.106:6801/285173716","cluster_addr":"192.168.123.106:6803/285173716","heartbeat_back_addr":"192.168.123.106:6807/285173716","heartbeat_front_addr":"192.168.123.106:6805/285173716","state":["exists","up"]},{"osd":3,"uuid":"e5eb8885-521e-4b6c-8370-7d8604264779","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":38,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6809","nonce":2984060640}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6811","nonce":2984060640}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6815","nonce":2984060640}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":2984060640},{"type":"v1","addr":"192.168.123.106:6813","nonce":2984060640}]},"public_addr":"192.168.123.106:6809/2984060640","cluster_addr":"192.168.123.106:6811/2984060640","heartbeat_back_addr":"192.168.123.106:6815/2984060640","heartbeat_front_addr":"192.168.123.106:6813/2984060640","state":["exists","up"]},{"osd":4,"uuid":"1c675835-f3b7-40fb-a62f-e795b68d3dcb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":31,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6801","nonce":726545420}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6803","nonce":726545420}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6807","nonce":726545420}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":726545420},{"type":"v1","addr":"192.168.123.109:6805","nonce":726545420}]},"public_addr":"192.168.123.109:6801/726545420","cluster_addr":"192.168.123.109:6803/726545420","heartbeat_back_addr":"192.168.123.109:6807/726545420","heartbeat_front_addr":"192.168.123.109:6805/726545420","state":["exists","up"]},{"osd":5,"uuid":"329c35de-29d1-40e7-bb6f-b4581bee34b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":37,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6809","nonce":3212082069}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6811","nonce":3212082069}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6815","nonce":3212082069}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3212082069},{"type":"v1","addr":"192.168.123.109:6813","nonce":3212082069}]},"public_addr":"192.168.123.109:6809/3212082069","cluster_addr":"192.168.123.109:6811/3212082069","heartbeat_back_addr":"192.168.123.109:6815/3212082069","heartbeat_front_addr":"192.168.123.109:6813/3212082069","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:15.734917+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:24.957450+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:34.284347+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:44.190420+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:58:53.436462+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T12:59:03.806044+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.104:6801/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/1226904767":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/2118987935":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6801/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/3807242158":"2026-03-11T12:57:34.270523+0000","192.168.123.104:6800/3093816275":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2934451579":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/1685098299":"2026-03-11T12:57:23.974125+0000","192.168.123.104:0/2928968844":"2026-03-11T12:57:34.270523+0000","192.168.123.104:0/3806698536":"2026-03-11T12:57:23.974125+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T12:59:09.547 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:09 vm04 ceph-mon[50620]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 560 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:09.547 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:09 vm04 ceph-mon[50620]: osdmap e39: 6 total, 6 up, 6 in 2026-03-10T12:59:09.547 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:09 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3286600337' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.547 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:09 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/343817636' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.547 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:09 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1767991728' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.0 flush_pg_stats 2026-03-10T12:59:09.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.1 flush_pg_stats 2026-03-10T12:59:09.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.2 flush_pg_stats 2026-03-10T12:59:09.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.3 flush_pg_stats 2026-03-10T12:59:09.548 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.4 flush_pg_stats 2026-03-10T12:59:09.549 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph tell osd.5 flush_pg_stats 2026-03-10T12:59:09.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:09 vm09 ceph-mon[53016]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 560 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:09.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:09 vm09 ceph-mon[53016]: osdmap e39: 6 total, 6 up, 6 in 2026-03-10T12:59:09.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:09 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3286600337' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:09 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/343817636' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:09 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1767991728' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:09 vm06 ceph-mon[56046]: pgmap v75: 1 pgs: 1 remapped+peering; 449 KiB data, 560 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:09.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:09 vm06 ceph-mon[56046]: osdmap e39: 6 total, 6 up, 6 in 2026-03-10T12:59:09.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:09 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3286600337' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:09 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/343817636' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:09.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:09 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1767991728' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T12:59:10.072 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.072 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.075 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.077 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.112 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.321 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:10.515 INFO:teuthology.orchestra.run.vm04.stdout:73014444040 2026-03-10T12:59:10.515 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.2 2026-03-10T12:59:10.660 INFO:teuthology.orchestra.run.vm04.stdout:133143986180 2026-03-10T12:59:10.660 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.4 2026-03-10T12:59:10.780 INFO:teuthology.orchestra.run.vm04.stdout:158913789954 2026-03-10T12:59:10.780 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.5 2026-03-10T12:59:10.836 INFO:teuthology.orchestra.run.vm04.stdout:51539607562 2026-03-10T12:59:10.836 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.1 2026-03-10T12:59:10.850 INFO:teuthology.orchestra.run.vm04.stdout:34359738380 2026-03-10T12:59:10.850 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.0 2026-03-10T12:59:10.857 INFO:teuthology.orchestra.run.vm04.stdout:107374182406 2026-03-10T12:59:10.857 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.3 2026-03-10T12:59:11.063 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.252 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.479 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.506 INFO:teuthology.orchestra.run.vm04.stdout:73014444039 2026-03-10T12:59:11.622 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444040 got 73014444039 for osd.2 2026-03-10T12:59:11.625 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.682 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.682 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:11.699 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:11 vm04 ceph-mon[50620]: pgmap v77: 1 pgs: 1 remapped+peering; 449 KiB data, 561 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:11.699 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:11 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3648257823' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:11.801 INFO:teuthology.orchestra.run.vm04.stdout:133143986179 2026-03-10T12:59:11.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:11 vm09 ceph-mon[53016]: pgmap v77: 1 pgs: 1 remapped+peering; 449 KiB data, 561 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:11.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:11 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3648257823' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:11.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:11 vm06 ceph-mon[56046]: pgmap v77: 1 pgs: 1 remapped+peering; 449 KiB data, 561 MiB used, 119 GiB / 120 GiB avail 2026-03-10T12:59:11.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:11 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3648257823' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:11.961 INFO:tasks.cephadm.ceph_manager.ceph:need seq 133143986180 got 133143986179 for osd.4 2026-03-10T12:59:12.077 INFO:teuthology.orchestra.run.vm04.stdout:34359738379 2026-03-10T12:59:12.116 INFO:teuthology.orchestra.run.vm04.stdout:51539607561 2026-03-10T12:59:12.177 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738380 got 34359738379 for osd.0 2026-03-10T12:59:12.201 INFO:teuthology.orchestra.run.vm04.stdout:158913789953 2026-03-10T12:59:12.224 INFO:teuthology.orchestra.run.vm04.stdout:107374182405 2026-03-10T12:59:12.231 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607562 got 51539607561 for osd.1 2026-03-10T12:59:12.285 INFO:tasks.cephadm.ceph_manager.ceph:need seq 158913789954 got 158913789953 for osd.5 2026-03-10T12:59:12.311 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182406 got 107374182405 for osd.3 2026-03-10T12:59:12.545 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:12 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2949620805' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:12.545 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:12 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2352492992' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:12.545 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:12 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3939416390' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:12.545 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:12 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1392503760' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:12.545 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:12 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/1187039396' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:12.623 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.2 2026-03-10T12:59:12.792 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:12.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:12 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2949620805' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:12.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:12 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2352492992' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:12.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:12 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3939416390' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:12.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:12 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1392503760' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:12.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:12 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/1187039396' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:12.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:12 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2949620805' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:12.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:12 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2352492992' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:12.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:12 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3939416390' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:12.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:12 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1392503760' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:12.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:12 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/1187039396' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:12.961 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.4 2026-03-10T12:59:13.025 INFO:teuthology.orchestra.run.vm04.stdout:73014444040 2026-03-10T12:59:13.095 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444040 got 73014444040 for osd.2 2026-03-10T12:59:13.095 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:13.157 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:13.178 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.0 2026-03-10T12:59:13.232 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.1 2026-03-10T12:59:13.286 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.5 2026-03-10T12:59:13.312 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph osd last-stat-seq osd.3 2026-03-10T12:59:13.450 INFO:teuthology.orchestra.run.vm04.stdout:133143986181 2026-03-10T12:59:13.567 INFO:tasks.cephadm.ceph_manager.ceph:need seq 133143986180 got 133143986181 for osd.4 2026-03-10T12:59:13.567 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:13.587 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:13.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:13 vm04 ceph-mon[50620]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:13.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:13 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2569305015' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:13.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:13 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2272796123' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:13.745 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:13 vm09 ceph-mon[53016]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:13 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2569305015' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:13.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:13 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2272796123' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:13.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:13 vm06 ceph-mon[56046]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-10T12:59:13.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:13 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2569305015' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T12:59:13.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:13 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2272796123' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T12:59:13.888 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:13.906 INFO:teuthology.orchestra.run.vm04.stdout:34359738380 2026-03-10T12:59:13.972 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:14.036 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738380 got 34359738380 for osd.0 2026-03-10T12:59:14.036 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:14.187 INFO:teuthology.orchestra.run.vm04.stdout:51539607563 2026-03-10T12:59:14.234 INFO:teuthology.orchestra.run.vm04.stdout:158913789954 2026-03-10T12:59:14.249 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607562 got 51539607563 for osd.1 2026-03-10T12:59:14.249 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:14.301 INFO:tasks.cephadm.ceph_manager.ceph:need seq 158913789954 got 158913789954 for osd.5 2026-03-10T12:59:14.301 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:14.357 INFO:teuthology.orchestra.run.vm04.stdout:107374182407 2026-03-10T12:59:14.405 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182406 got 107374182407 for osd.3 2026-03-10T12:59:14.405 DEBUG:teuthology.parallel:result is None 2026-03-10T12:59:14.405 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T12:59:14.405 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph pg dump --format=json 2026-03-10T12:59:14.577 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:14.791 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/203160496' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:14.791 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3095603203' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:14.791 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/2837168141' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:14.791 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:14 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3198823787' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:14.792 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:14.792 INFO:teuthology.orchestra.run.vm04.stderr:dumped all 2026-03-10T12:59:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:14 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/203160496' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:14 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3095603203' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:14 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/2837168141' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:14.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:14 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3198823787' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:14.854 INFO:teuthology.orchestra.run.vm04.stdout:{"pg_ready":true,"pg_map":{"version":79,"stamp":"2026-03-10T12:59:14.297808+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":3,"kb":125804544,"kb_used":164544,"kb_used_data":3492,"kb_used_omap":9,"kb_used_meta":160886,"kb_avail":125640000,"statfs":{"total":128823853056,"available":128655360000,"internally_reserved":0,"allocated":3575808,"data_stored":2199636,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9538,"internal_metadata":164747966},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"7.789147"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":59,"reported_epoch":39,"state":"active+clean","last_fresh":"2026-03-10T12:59:08.756630+0000","last_change":"2026-03-10T12:59:08.756630+0000","last_active":"2026-03-10T12:59:08.756630+0000","last_peered":"2026-03-10T12:59:08.756630+0000","last_clean":"2026-03-10T12:59:08.756630+0000","last_became_active":"2026-03-10T12:59:08.751277+0000","last_became_peered":"2026-03-10T12:59:08.751277+0000","last_unstale":"2026-03-10T12:59:08.756630+0000","last_undegraded":"2026-03-10T12:59:08.756630+0000","last_fullsized":"2026-03-10T12:59:08.756630+0000","mapping_epoch":38,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":39,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T12:58:39.252019+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T12:58:39.252019+0000","last_clean_scrub_stamp":"2026-03-10T12:58:39.252019+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-12T00:13:47.389632+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":6}],"osd_stats":[{"osd":5,"up_from":37,"seq":158913789955,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":31,"seq":133143986181,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27192,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940232,"statfs":{"total":21470642176,"available":21442797568,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":25,"seq":107374182407,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":17,"seq":73014444041,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607563,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27200,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940224,"statfs":{"total":21470642176,"available":21442789376,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738380,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27200,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940224,"statfs":{"total":21470642176,"available":21442789376,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T12:59:14.854 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph pg dump --format=json 2026-03-10T12:59:14.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:14 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/203160496' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T12:59:14.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:14 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3095603203' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T12:59:14.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:14 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/2837168141' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T12:59:14.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:14 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3198823787' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T12:59:15.017 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:15.227 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:15.227 INFO:teuthology.orchestra.run.vm04.stderr:dumped all 2026-03-10T12:59:15.274 INFO:teuthology.orchestra.run.vm04.stdout:{"pg_ready":true,"pg_map":{"version":79,"stamp":"2026-03-10T12:59:14.297808+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":3,"kb":125804544,"kb_used":164544,"kb_used_data":3492,"kb_used_omap":9,"kb_used_meta":160886,"kb_avail":125640000,"statfs":{"total":128823853056,"available":128655360000,"internally_reserved":0,"allocated":3575808,"data_stored":2199636,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9538,"internal_metadata":164747966},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"7.789147"},"pg_stats":[{"pgid":"1.0","version":"20'32","reported_seq":59,"reported_epoch":39,"state":"active+clean","last_fresh":"2026-03-10T12:59:08.756630+0000","last_change":"2026-03-10T12:59:08.756630+0000","last_active":"2026-03-10T12:59:08.756630+0000","last_peered":"2026-03-10T12:59:08.756630+0000","last_clean":"2026-03-10T12:59:08.756630+0000","last_became_active":"2026-03-10T12:59:08.751277+0000","last_became_peered":"2026-03-10T12:59:08.751277+0000","last_unstale":"2026-03-10T12:59:08.756630+0000","last_undegraded":"2026-03-10T12:59:08.756630+0000","last_fullsized":"2026-03-10T12:59:08.756630+0000","mapping_epoch":38,"log_start":"0'0","ondisk_log_start":"0'0","created":19,"last_epoch_clean":39,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T12:58:39.252019+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T12:58:39.252019+0000","last_clean_scrub_stamp":"2026-03-10T12:58:39.252019+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-12T00:13:47.389632+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":6}],"osd_stats":[{"osd":5,"up_from":37,"seq":158913789955,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":31,"seq":133143986181,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27192,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940232,"statfs":{"total":21470642176,"available":21442797568,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":25,"seq":107374182407,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27648,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939776,"statfs":{"total":21470642176,"available":21442330624,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":17,"seq":73014444041,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27652,"kb_used_data":808,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939772,"statfs":{"total":21470642176,"available":21442326528,"internally_reserved":0,"allocated":827392,"data_stored":596246,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":12,"seq":51539607563,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27200,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940224,"statfs":{"total":21470642176,"available":21442789376,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738380,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27200,"kb_used_data":356,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940224,"statfs":{"total":21470642176,"available":21442789376,"internally_reserved":0,"allocated":364544,"data_stored":136966,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T12:59:15.274 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T12:59:15.274 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T12:59:15.274 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T12:59:15.274 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph health --format=json 2026-03-10T12:59:15.439 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:15.679 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T12:59:15.679 INFO:teuthology.orchestra.run.vm04.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T12:59:15.725 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:15 vm04 ceph-mon[50620]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T12:59:15.726 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:15 vm04 ceph-mon[50620]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:15.726 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T12:59:15.726 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T12:59:15.726 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T12:59:15.729 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm04.local 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- bash -c 'set -ex 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> HOSTNAMES=$(ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"') 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> for host in $HOSTNAMES; do 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # find the hostname for "host.c" which will have no mgr 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> HAS_MGRS=$(ceph orch ps --hostname ${host} --format json | jq '"'"'any(.daemon_type == "mgr")'"'"') 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> if [ "$HAS_MGRS" == "false" ]; then 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> HOST_C="${host}" 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> fi 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # One last thing to worry about before draining the host 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # is that the teuthology test tends to put the explicit 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # hostnames in the placement for the mon service. 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # We want to make sure we can drain without providing 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # --force and there is a check for the host being removed 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # being listed explicitly in the placements. Therefore, 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # we should remove it from the mon placement. 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> ceph orch ls mon --export > mon.yaml 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> sed /"$HOST_C"/d mon.yaml > mon_adjusted.yaml 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> ceph orch apply -i mon_adjusted.yaml 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # now drain that host 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> ceph orch host drain $HOST_C --zap-osd-devices 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # wait for drain to complete 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C) 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> while [ "$HOST_C_DAEMONS" != "No daemons reported" ]; do 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> sleep 15 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C) 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # we want to check the ability to remove the host from 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # the CRUSH map, so we should first verify the host is in 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # the CRUSH map. 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> ceph osd getcrushmap -o compiled-crushmap 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> crushtool -d compiled-crushmap -o crushmap.txt 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> CRUSH_MAP=$(cat crushmap.txt) 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> if ! grep -q "$HOST_C" <<< "$CRUSH_MAP"; then 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> printf "Expected to see $HOST_C in CRUSH map. Saw:\n\n$CRUSH_MAP" 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> exit 1 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> fi 2026-03-10T12:59:15.730 DEBUG:teuthology.orchestra.run.vm04:> # If the drain was successful, we should be able to remove the 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> # host without force with no issues. If there are still daemons 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> # we will get a response telling us to drain the host and a 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> # non-zero return code 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> ceph orch host rm $HOST_C --rm-crush-entry 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> # verify we'"'"'ve successfully removed the host from the CRUSH map 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> sleep 30 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> ceph osd getcrushmap -o compiled-crushmap 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> crushtool -d compiled-crushmap -o crushmap.txt 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> CRUSH_MAP=$(cat crushmap.txt) 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> if grep -q "$HOST_C" <<< "$CRUSH_MAP"; then 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> printf "Saw $HOST_C in CRUSH map after it should have been removed.\n\n$CRUSH_MAP" 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> exit 1 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> fi 2026-03-10T12:59:15.731 DEBUG:teuthology.orchestra.run.vm04:> ' 2026-03-10T12:59:15.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:15 vm09 ceph-mon[53016]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T12:59:15.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:15 vm09 ceph-mon[53016]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:15.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:15 vm06 ceph-mon[56046]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 58 KiB/s, 0 objects/s recovering 2026-03-10T12:59:15.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:15 vm06 ceph-mon[56046]: from='client.14508 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:15.902 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T12:59:15.972 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch host ls --format json 2026-03-10T12:59:15.976 INFO:teuthology.orchestra.run.vm04.stderr:++ jq -r '.[] | .hostname' 2026-03-10T12:59:16.131 INFO:teuthology.orchestra.run.vm04.stderr:+ HOSTNAMES='vm04 2026-03-10T12:59:16.131 INFO:teuthology.orchestra.run.vm04.stderr:vm06 2026-03-10T12:59:16.131 INFO:teuthology.orchestra.run.vm04.stderr:vm09' 2026-03-10T12:59:16.131 INFO:teuthology.orchestra.run.vm04.stderr:+ for host in $HOSTNAMES 2026-03-10T12:59:16.132 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm04 --format json 2026-03-10T12:59:16.132 INFO:teuthology.orchestra.run.vm04.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-10T12:59:16.301 INFO:teuthology.orchestra.run.vm04.stderr:+ HAS_MGRS=true 2026-03-10T12:59:16.301 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' true == false ']' 2026-03-10T12:59:16.301 INFO:teuthology.orchestra.run.vm04.stderr:+ for host in $HOSTNAMES 2026-03-10T12:59:16.301 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm06 --format json 2026-03-10T12:59:16.301 INFO:teuthology.orchestra.run.vm04.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-10T12:59:16.466 INFO:teuthology.orchestra.run.vm04.stderr:+ HAS_MGRS=true 2026-03-10T12:59:16.466 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' true == false ']' 2026-03-10T12:59:16.466 INFO:teuthology.orchestra.run.vm04.stderr:+ for host in $HOSTNAMES 2026-03-10T12:59:16.467 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm09 --format json 2026-03-10T12:59:16.467 INFO:teuthology.orchestra.run.vm04.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-10T12:59:16.568 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:16 vm04 ceph-mon[50620]: from='client.14514 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.569 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:16 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/3009590539' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:59:16.569 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:16 vm04 ceph-mon[50620]: from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.641 INFO:teuthology.orchestra.run.vm04.stderr:+ HAS_MGRS=false 2026-03-10T12:59:16.641 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' false == false ']' 2026-03-10T12:59:16.641 INFO:teuthology.orchestra.run.vm04.stderr:+ HOST_C=vm09 2026-03-10T12:59:16.641 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph orch ls mon --export 2026-03-10T12:59:16.807 INFO:teuthology.orchestra.run.vm04.stderr:+ sed /vm09/d mon.yaml 2026-03-10T12:59:16.808 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph orch apply -i mon_adjusted.yaml 2026-03-10T12:59:16.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:16 vm09 ceph-mon[53016]: from='client.14514 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:16 vm09 ceph-mon[53016]: from='client.? 192.168.123.104:0/3009590539' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:59:16.847 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:16 vm09 ceph-mon[53016]: from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:16 vm06 ceph-mon[56046]: from='client.14514 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:16 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/3009590539' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T12:59:16.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:16 vm06 ceph-mon[56046]: from='client.14526 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:16.973 INFO:teuthology.orchestra.run.vm04.stdout:Scheduled mon update... 2026-03-10T12:59:16.995 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph orch host drain vm09 --zap-osd-devices 2026-03-10T12:59:17.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:17 vm09 podman[64654]: 2026-03-10 12:59:17.012176951 +0000 UTC m=+0.010305161 container died 5d24d3ec9650c0820465268be8fac379d63ec5c937e694cdea212180846772fc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-c, ceph=True, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T12:59:17.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:17 vm09 podman[64654]: 2026-03-10 12:59:17.023803365 +0000 UTC m=+0.021931566 container remove 5d24d3ec9650c0820465268be8fac379d63ec5c937e694cdea212180846772fc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-c, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, OSD_FLAVOR=default, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T12:59:17.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:17 vm09 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.c.service: Deactivated successfully. 2026-03-10T12:59:17.348 INFO:journalctl@ceph.mon.c.vm09.stdout:Mar 10 12:59:17 vm09 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.c.service: Consumed 1.347s CPU time. 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: purged_snaps scrub ok 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm04", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 51 KiB/s, 0 objects/s recovering 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm06", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_type": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='client.14556 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: Saving service mon spec with placement vm04:192.168.123.104=a;vm06:192.168.123.106=b;count:3 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: Safe to remove mon.c: new quorum should be ['a', 'b'] (from ['a', 'b']) 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: Removing monitor c from monmap... 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "mon rm", "name": "c"}]': finished 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: mon.b calling monitor election 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: mon.a calling monitor election 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: monmap epoch 4 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: last_changed 2026-03-10T12:59:16.979779+0000 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: min_mon_release 19 (squid) 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: election_strategy: 1 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: fsmap 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: osdmap e39: 6 total, 6 up, 6 in 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: mgrmap e14: a(active, since 104s), standbys: b 2026-03-10T12:59:19.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:19 vm06 ceph-mon[56046]: overall HEALTH_OK 2026-03-10T12:59:19.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: purged_snaps scrub ok 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='client.14532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm04", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 51 KiB/s, 0 objects/s recovering 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm06", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='client.14544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_type": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='client.14556 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: Saving service mon spec with placement vm04:192.168.123.104=a;vm06:192.168.123.106=b;count:3 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: Safe to remove mon.c: new quorum should be ['a', 'b'] (from ['a', 'b']) 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: Removing monitor c from monmap... 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "mon rm", "name": "c"}]': finished 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: mon.b calling monitor election 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: mon.a calling monitor election 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: monmap epoch 4 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: last_changed 2026-03-10T12:59:16.979779+0000 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: created 2026-03-10T12:57:11.405959+0000 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: min_mon_release 19 (squid) 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: election_strategy: 1 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: 0: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.a 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: 1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: fsmap 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: osdmap e39: 6 total, 6 up, 6 in 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: mgrmap e14: a(active, since 104s), standbys: b 2026-03-10T12:59:19.378 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:19 vm04 ceph-mon[50620]: overall HEALTH_OK 2026-03-10T12:59:20.223 INFO:teuthology.orchestra.run.vm04.stdout:Scheduled to remove the following daemons from host 'vm09' 2026-03-10T12:59:20.223 INFO:teuthology.orchestra.run.vm04.stdout:type id 2026-03-10T12:59:20.223 INFO:teuthology.orchestra.run.vm04.stdout:-------------------- --------------- 2026-03-10T12:59:20.223 INFO:teuthology.orchestra.run.vm04.stdout:osd 4 2026-03-10T12:59:20.223 INFO:teuthology.orchestra.run.vm04.stdout:osd 5 2026-03-10T12:59:20.233 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm09 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: Removing daemon mon.c from vm09 -- ports [] 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:20.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:20 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: Removing daemon mon.c from vm09 -- ports [] 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:20.374 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:20 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:+ HOST_C_DAEMONS='NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:osd.4 vm09 running (29s) 0s ago 29s 58.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 5b2e0b5706de 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:osd.5 vm09 running (19s) 0s ago 18s 56.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 1dda985ae84d ' 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' 'NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:osd.4 vm09 running (29s) 0s ago 29s 58.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 5b2e0b5706de 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:osd.5 vm09 running (19s) 0s ago 18s 56.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 1dda985ae84d ' '!=' 'No daemons reported' ']' 2026-03-10T12:59:20.426 INFO:teuthology.orchestra.run.vm04.stderr:+ sleep 15 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.192 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Reconfiguring mon.a (monmap changed)... 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: Reconfiguring daemon mon.a on vm04 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:59:21.193 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:21 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm04:/etc/ceph/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Updating vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Reconfiguring mon.a (monmap changed)... 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: Reconfiguring daemon mon.a on vm04 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T12:59:21.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:21 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='client.24397 -' entity='client.admin' cmd=[{"prefix": "orch host drain", "hostname": "vm09", "zap_osd_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Added label _no_schedule to host vm09 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Added label _no_conf_keyring to host vm09 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: osd.4 crush weight is 0.0194854736328125 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: osd.5 crush weight is 0.0194854736328125 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 38 KiB/s, 0 objects/s recovering 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='client.14564 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring mgr.a (monmap changed)... 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring daemon mgr.a on vm04 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring osd.0 (monmap changed)... 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring daemon osd.0 on vm04 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring osd.1 (monmap changed)... 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: Reconfiguring daemon osd.1 on vm04 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.321 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T12:59:22.322 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:22 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='client.24397 -' entity='client.admin' cmd=[{"prefix": "orch host drain", "hostname": "vm09", "zap_osd_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Added label _no_schedule to host vm09 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Added label _no_conf_keyring to host vm09 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: osd.4 crush weight is 0.0194854736328125 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: osd.5 crush weight is 0.0194854736328125 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 38 KiB/s, 0 objects/s recovering 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='client.14564 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring mgr.a (monmap changed)... 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring daemon mgr.a on vm04 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring osd.0 (monmap changed)... 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring daemon osd.0 on vm04 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring osd.1 (monmap changed)... 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: Reconfiguring daemon osd.1 on vm04 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T12:59:22.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:22 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring mon.b (monmap changed)... 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring daemon mon.b on vm06 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring mgr.b (monmap changed)... 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring daemon mgr.b on vm06 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring osd.2 (monmap changed)... 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring daemon osd.2 on vm06 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring osd.3 (monmap changed)... 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: Reconfiguring daemon osd.3 on vm06 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:23.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-10T12:59:23.611 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:23 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring mon.b (monmap changed)... 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring daemon mon.b on vm06 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring mgr.b (monmap changed)... 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring daemon mgr.b on vm06 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring osd.2 (monmap changed)... 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring daemon osd.2 on vm06 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring osd.3 (monmap changed)... 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: Reconfiguring daemon osd.3 on vm06 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-10T12:59:23.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:23 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a[50616]: 2026-03-10T12:59:24.112+0000 7fd667a7a640 -1 mon.a@0(leader).osd e40 definitely_dead 0 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Reconfiguring osd.4 (monmap changed)... 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Reconfiguring daemon osd.4 on vm09 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Reconfiguring osd.5 (monmap changed)... 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Reconfiguring daemon osd.5 on vm09 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Removing vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Removing vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: Removing vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.client.admin.keyring 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]': finished 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: osdmap e40: 6 total, 6 up, 6 in 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-10T12:59:24.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:24 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["4"]}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Reconfiguring osd.4 (monmap changed)... 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Reconfiguring daemon osd.4 on vm09 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Reconfiguring osd.5 (monmap changed)... 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Reconfiguring daemon osd.5 on vm09 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Removing vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.conf 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Removing vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: Removing vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/config/ceph.client.admin.keyring 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]': finished 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: osdmap e40: 6 total, 6 up, 6 in 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-10T12:59:24.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:24 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["4"]}]: dispatch 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 systemd[1]: Stopping Ceph osd.4 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:59:25.330+0000 7f6274b88640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:59:25.330+0000 7f6274b88640 -1 osd.4 40 *** Got signal Terminated *** 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4[56500]: 2026-03-10T12:59:25.330+0000 7f6274b88640 -1 osd.4 40 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65652]: 2026-03-10 12:59:25.362008859 +0000 UTC m=+0.043541238 container died 5b2e0b5706dee1935726817b587d29a3b534c266c9b937206a3a98781a8b0cdc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4, CEPH_REF=squid, ceph=True, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default) 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65652]: 2026-03-10 12:59:25.389432308 +0000 UTC m=+0.070964667 container remove 5b2e0b5706dee1935726817b587d29a3b534c266c9b937206a3a98781a8b0cdc (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, ceph=True, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T12:59:25.543 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 bash[65652]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: osd.5 weight is now 0.0 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["4"]}]': finished 2026-03-10T12:59:25.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:25 vm06 ceph-mon[56046]: osdmap e41: 6 total, 5 up, 6 in 2026-03-10T12:59:25.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: osd.5 weight is now 0.0 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["4"]}]': finished 2026-03-10T12:59:25.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:25 vm04 ceph-mon[50620]: osdmap e41: 6 total, 5 up, 6 in 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.541620693 +0000 UTC m=+0.022644690 container create e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.585021016 +0000 UTC m=+0.066045003 container init e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, ceph=True, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.591730351 +0000 UTC m=+0.072754338 container start e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.592672005 +0000 UTC m=+0.073695992 container attach e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.build-date=20260223, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.529984391 +0000 UTC m=+0.011008397 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.721753645 +0000 UTC m=+0.202777642 container died e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, org.label-schema.schema-version=1.0, ceph=True, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 podman[65717]: 2026-03-10 12:59:25.737904222 +0000 UTC m=+0.218928209 container remove e2dd6e0a53e85102bd8b0723a13c565100631e3422ff17b267b933c67a5150df (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-4-deactivate, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.4.service: Deactivated successfully. 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.4.service: Unit process 65728 (conmon) remains running after unit stopped. 2026-03-10T12:59:25.816 INFO:journalctl@ceph.osd.4.vm09.stdout:Mar 10 12:59:25 vm09 systemd[1]: Stopped Ceph osd.4 for 74d0e8a6-1c80-11f1-9542-4d7926099318. 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: osd.4 now down 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: Removing daemon osd.4 from vm09 -- ports [] 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.4"}]: dispatch 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.4"}]': finished 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: Cluster is now healthy 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:59:26.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:26 vm06 ceph-mon[56046]: osdmap e42: 5 total, 5 up, 5 in 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: osd.4 now down 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: Removing daemon osd.4 from vm09 -- ports [] 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.4"}]: dispatch 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.4"}]': finished 2026-03-10T12:59:26.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:59:26.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-10T12:59:26.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: Cluster is now healthy 2026-03-10T12:59:26.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:59:26.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:26 vm04 ceph-mon[50620]: osdmap e42: 5 total, 5 up, 5 in 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: Removing key for osd.4 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: Successfully removed osd.4 on vm09 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: Successfully purged osd.4 on vm09 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: Zapping devices for osd.4 on vm09 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: osdmap e43: 5 total, 5 up, 5 in 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: Successfully zapped devices for osd.4 on vm09 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:27.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:27 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: Removing key for osd.4 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: Successfully removed osd.4 on vm09 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: Successfully purged osd.4 on vm09 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: Zapping devices for osd.4 on vm09 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: osdmap e43: 5 total, 5 up, 5 in 2026-03-10T12:59:27.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: Successfully zapped devices for osd.4 on vm09 2026-03-10T12:59:27.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:27.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:27.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:27.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:27 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: osdmap e44: 5 total, 5 up, 5 in 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:28 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:28.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: osdmap e44: 5 total, 5 up, 5 in 2026-03-10T12:59:28.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:28.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:28 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: Detected new or changed devices on vm09 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: pgmap v91: 1 pgs: 1 remapped+peering; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:29 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:29.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: Detected new or changed devices on vm09 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: pgmap v91: 1 pgs: 1 remapped+peering; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:29.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:29 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:30.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:30 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:30.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:30 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:30.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:30 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:30.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:30 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:30.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:30 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:30.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:30 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:31.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:31 vm06 ceph-mon[56046]: pgmap v92: 1 pgs: 1 remapped+peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:31.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:31 vm04 ceph-mon[50620]: pgmap v92: 1 pgs: 1 remapped+peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-10T12:59:33.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:33 vm06 ceph-mon[56046]: pgmap v93: 1 pgs: 1 active+clean+wait; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail; 71 KiB/s, 0 objects/s recovering 2026-03-10T12:59:33.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:33 vm04 ceph-mon[50620]: pgmap v93: 1 pgs: 1 active+clean+wait; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail; 71 KiB/s, 0 objects/s recovering 2026-03-10T12:59:34.627 INFO:journalctl@ceph.mgr.a.vm04.stdout:Mar 10 12:59:34 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mgr-a[50829]: 2026-03-10T12:59:34.293+0000 7f2d55049640 -1 calc_pg_upmaps abort due to max <= 0 2026-03-10T12:59:35.426 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm09 2026-03-10T12:59:35.595 INFO:teuthology.orchestra.run.vm04.stderr:+ HOST_C_DAEMONS='NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T12:59:35.595 INFO:teuthology.orchestra.run.vm04.stderr:osd.5 vm09 running (34s) 8s ago 34s 58.0M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 1dda985ae84d ' 2026-03-10T12:59:35.595 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' 'NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T12:59:35.596 INFO:teuthology.orchestra.run.vm04.stderr:osd.5 vm09 running (34s) 8s ago 34s 58.0M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 1dda985ae84d ' '!=' 'No daemons reported' ']' 2026-03-10T12:59:35.596 INFO:teuthology.orchestra.run.vm04.stderr:+ sleep 15 2026-03-10T12:59:35.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:35 vm06 ceph-mon[56046]: pgmap v94: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-10T12:59:35.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:35 vm04 ceph-mon[50620]: pgmap v94: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-10T12:59:36.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:35 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a[50616]: 2026-03-10T12:59:35.948+0000 7fd667a7a640 -1 mon.a@0(leader).osd e44 definitely_dead 0 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='client.14568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:36 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='client.14568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:36.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:36 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["5"]}]: dispatch 2026-03-10T12:59:36.877 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 systemd[1]: Stopping Ceph osd.5 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T12:59:36.877 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:36.689+0000 7f2559cd5640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T12:59:36.877 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:36.689+0000 7f2559cd5640 -1 osd.5 44 *** Got signal Terminated *** 2026-03-10T12:59:36.877 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5[61733]: 2026-03-10T12:59:36.689+0000 7f2559cd5640 -1 osd.5 44 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T12:59:36.877 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70128]: 2026-03-10 12:59:36.724120884 +0000 UTC m=+0.048082563 container died 1dda985ae84d310206ea0b89773d16ed108b2372d9ba2ec539e0dc6c43ce01e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, io.buildah.version=1.41.3, ceph=True, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-10T12:59:36.878 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70128]: 2026-03-10 12:59:36.743567354 +0000 UTC m=+0.067529024 container remove 1dda985ae84d310206ea0b89773d16ed108b2372d9ba2ec539e0dc6c43ce01e9 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223) 2026-03-10T12:59:36.878 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 bash[70128]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70195]: 2026-03-10 12:59:36.876102743 +0000 UTC m=+0.015119918 container create ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_REF=squid, ceph=True, io.buildah.version=1.41.3, OSD_FLAVOR=default) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70195]: 2026-03-10 12:59:36.918852025 +0000 UTC m=+0.057869210 container init ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70195]: 2026-03-10 12:59:36.921553875 +0000 UTC m=+0.060571050 container start ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70195]: 2026-03-10 12:59:36.926189788 +0000 UTC m=+0.065206963 container attach ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:36 vm09 podman[70195]: 2026-03-10 12:59:36.870197984 +0000 UTC m=+0.009215168 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:37 vm09 podman[70195]: 2026-03-10 12:59:37.04116601 +0000 UTC m=+0.180183175 container died ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, ceph=True) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:37 vm09 podman[70195]: 2026-03-10 12:59:37.053858892 +0000 UTC m=+0.192876067 container remove ded43f4a521a41d020dcc0f6c10d2e0b19601157282d6793f1fb7017c1cbe687 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-5-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:37 vm09 systemd[1]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.5.service: Deactivated successfully. 2026-03-10T12:59:37.142 INFO:journalctl@ceph.osd.5.vm09.stdout:Mar 10 12:59:37 vm09 systemd[1]: Stopped Ceph osd.5 for 74d0e8a6-1c80-11f1-9542-4d7926099318. 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: pgmap v95: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 48 KiB/s, 0 objects/s recovering 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["5"]}]': finished 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: osdmap e45: 5 total, 4 up, 5 in 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: osd.5 now down 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Removing daemon osd.5 from vm09 -- ports [] 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.5"}]': finished 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: Cluster is now healthy 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:59:37.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:37 vm06 ceph-mon[56046]: osdmap e46: 4 total, 4 up, 4 in 2026-03-10T12:59:37.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: pgmap v95: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 48 KiB/s, 0 objects/s recovering 2026-03-10T12:59:37.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-10T12:59:37.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["5"]}]': finished 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: osdmap e45: 5 total, 4 up, 5 in 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: osd.5 now down 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Removing daemon osd.5 from vm09 -- ports [] 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.5"}]: dispatch 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.5"}]': finished 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: Cluster is now healthy 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]': finished 2026-03-10T12:59:37.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:37 vm04 ceph-mon[50620]: osdmap e46: 4 total, 4 up, 4 in 2026-03-10T12:59:38.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:38 vm06 ceph-mon[56046]: Removing key for osd.5 2026-03-10T12:59:38.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:38 vm06 ceph-mon[56046]: Successfully removed osd.5 on vm09 2026-03-10T12:59:38.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:38 vm06 ceph-mon[56046]: Successfully purged osd.5 on vm09 2026-03-10T12:59:38.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:38 vm06 ceph-mon[56046]: Zapping devices for osd.5 on vm09 2026-03-10T12:59:38.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:38 vm04 ceph-mon[50620]: Removing key for osd.5 2026-03-10T12:59:38.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:38 vm04 ceph-mon[50620]: Successfully removed osd.5 on vm09 2026-03-10T12:59:38.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:38 vm04 ceph-mon[50620]: Successfully purged osd.5 on vm09 2026-03-10T12:59:38.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:38 vm04 ceph-mon[50620]: Zapping devices for osd.5 on vm09 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: Successfully zapped devices for osd.5 on vm09 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:39.860 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:39 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:39.876 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-10T12:59:39.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: Successfully zapped devices for osd.5 on vm09 2026-03-10T12:59:39.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:39.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:39.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:39.877 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:39 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: Detected new or changed devices on vm09 2026-03-10T12:59:41.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.126 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:41.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:41.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.127 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:40 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: Detected new or changed devices on vm09 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:41.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:40 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:42.360 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:42 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:42.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:42.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:42.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:42.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:42.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:42.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:42 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:44.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:44 vm04 ceph-mon[50620]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:44.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:44 vm06 ceph-mon[56046]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:46.377 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:46 vm04 ceph-mon[50620]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:46.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:46 vm06 ceph-mon[56046]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:48.376 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:48 vm04 ceph-mon[50620]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:48.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:48 vm06 ceph-mon[56046]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:50.598 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:50 vm04 ceph-mon[50620]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:50.598 INFO:teuthology.orchestra.run.vm04.stderr:++ ceph orch ps --hostname vm09 2026-03-10T12:59:50.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:50 vm06 ceph-mon[56046]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:50.762 INFO:teuthology.orchestra.run.vm04.stderr:+ HOST_C_DAEMONS='No daemons reported' 2026-03-10T12:59:50.762 INFO:teuthology.orchestra.run.vm04.stderr:+ '[' 'No daemons reported' '!=' 'No daemons reported' ']' 2026-03-10T12:59:50.762 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph osd getcrushmap -o compiled-crushmap 2026-03-10T12:59:50.908 INFO:teuthology.orchestra.run.vm04.stderr:17 2026-03-10T12:59:50.918 INFO:teuthology.orchestra.run.vm04.stderr:+ crushtool -d compiled-crushmap -o crushmap.txt 2026-03-10T12:59:50.928 INFO:teuthology.orchestra.run.vm04.stderr:++ cat crushmap.txt 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:+ CRUSH_MAP='# begin crush map 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_local_tries 0 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_local_fallback_tries 0 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_total_tries 50 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_descend_once 1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_vary_r 1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_stable 1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable straw_calc_version 1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:tunable allowed_bucket_algs 54 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:# devices 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:device 0 osd.0 class hdd 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:device 1 osd.1 class hdd 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:device 2 osd.2 class hdd 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:device 3 osd.3 class hdd 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:# types 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 0 osd 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 1 host 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 2 chassis 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 3 rack 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 4 row 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 5 pdu 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 6 pod 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 7 room 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 8 datacenter 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 9 zone 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 10 region 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:type 11 root 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:# buckets 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:host vm04 { 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -3 # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -4 class hdd # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.03897 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: item osd.0 weight 0.01949 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: item osd.1 weight 0.01949 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:host vm06 { 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -5 # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -6 class hdd # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.03897 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: item osd.2 weight 0.01949 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: item osd.3 weight 0.01949 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:host vm09 { 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -7 # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: id -8 class hdd # do not change unnecessarily 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.00000 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T12:59:50.931 INFO:teuthology.orchestra.run.vm04.stderr:root default { 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: id -1 # do not change unnecessarily 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: id -2 class hdd # do not change unnecessarily 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.07794 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: item vm04 weight 0.03897 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: item vm06 weight 0.03897 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: item vm09 weight 0.00000 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:# rules 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:rule replicated_rule { 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: id 0 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: type replicated 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: step take default 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: step choose firstn 0 type osd 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: step emit 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:# end crush map' 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:+ grep -q vm09 2026-03-10T12:59:50.932 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph orch host rm vm09 --rm-crush-entry 2026-03-10T12:59:51.143 INFO:teuthology.orchestra.run.vm04.stdout:Removed host 'vm09' 2026-03-10T12:59:51.156 INFO:teuthology.orchestra.run.vm04.stderr:+ sleep 30 2026-03-10T12:59:51.379 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:51 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/4124527068' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-10T12:59:51.379 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:51 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush remove", "name": "vm09"}]: dispatch 2026-03-10T12:59:51.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:51 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/4124527068' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-10T12:59:51.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:51 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "osd crush remove", "name": "vm09"}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='client.14572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='client.14580 -' entity='client.admin' cmd=[{"prefix": "orch host rm", "hostname": "vm09", "rm_crush_entry": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd crush remove", "name": "vm09"}]': finished 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: osdmap e47: 4 total, 4 up, 4 in 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix":"config-key del","key":"mgr/cephadm/host.vm09"}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.vm09"}]': finished 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: Removed host vm09 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:52.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:52 vm06 ceph-mon[56046]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='client.14572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='client.14580 -' entity='client.admin' cmd=[{"prefix": "orch host rm", "hostname": "vm09", "rm_crush_entry": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix": "osd crush remove", "name": "vm09"}]': finished 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: osdmap e47: 4 total, 4 up, 4 in 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:52.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix":"config-key del","key":"mgr/cephadm/host.vm09"}]: dispatch 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.vm09"}]': finished 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: Removed host vm09 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T12:59:52.627 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:52 vm04 ceph-mon[50620]: from='mgr.14150 192.168.123.104:0/2316874078' entity='mgr.a' 2026-03-10T12:59:54.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:54 vm06 ceph-mon[56046]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:54.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:54 vm04 ceph-mon[50620]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:56.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:56 vm06 ceph-mon[56046]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:56.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:56 vm04 ceph-mon[50620]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:58.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 12:59:58 vm06 ceph-mon[56046]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T12:59:58.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 12:59:58 vm04 ceph-mon[50620]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:00.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:00 vm06 ceph-mon[56046]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:00.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:00 vm06 ceph-mon[56046]: overall HEALTH_OK 2026-03-10T13:00:00.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:00 vm04 ceph-mon[50620]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:00.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:00 vm04 ceph-mon[50620]: overall HEALTH_OK 2026-03-10T13:00:02.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:02 vm06 ceph-mon[56046]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:02.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:02 vm04 ceph-mon[50620]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:04.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:04 vm06 ceph-mon[56046]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:04.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:04 vm04 ceph-mon[50620]: pgmap v111: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:06.610 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:06 vm06 ceph-mon[56046]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:06.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:06 vm04 ceph-mon[50620]: pgmap v112: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:08.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:08 vm06 ceph-mon[56046]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:08.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:08 vm04 ceph-mon[50620]: pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:10.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:10 vm06 ceph-mon[56046]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:10.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:10 vm04 ceph-mon[50620]: pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:11.609 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:11 vm06 ceph-mon[56046]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:11.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:11 vm04 ceph-mon[50620]: pgmap v115: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:13.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:13 vm04 ceph-mon[50620]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:13.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:13 vm06 ceph-mon[56046]: pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:15.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:15 vm04 ceph-mon[50620]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:15.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:15 vm06 ceph-mon[56046]: pgmap v117: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:17.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:17 vm04 ceph-mon[50620]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:17.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:17 vm06 ceph-mon[56046]: pgmap v118: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:19.626 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:19 vm04 ceph-mon[50620]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:19.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:19 vm06 ceph-mon[56046]: pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:21.158 INFO:teuthology.orchestra.run.vm04.stderr:+ ceph osd getcrushmap -o compiled-crushmap 2026-03-10T13:00:21.308 INFO:teuthology.orchestra.run.vm04.stderr:18 2026-03-10T13:00:21.318 INFO:teuthology.orchestra.run.vm04.stderr:+ crushtool -d compiled-crushmap -o crushmap.txt 2026-03-10T13:00:21.327 INFO:teuthology.orchestra.run.vm04.stderr:++ cat crushmap.txt 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:+ CRUSH_MAP='# begin crush map 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_local_tries 0 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_local_fallback_tries 0 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable choose_total_tries 50 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_descend_once 1 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_vary_r 1 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable chooseleaf_stable 1 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable straw_calc_version 1 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:tunable allowed_bucket_algs 54 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:# devices 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:device 0 osd.0 class hdd 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:device 1 osd.1 class hdd 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:device 2 osd.2 class hdd 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:device 3 osd.3 class hdd 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:# types 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:type 0 osd 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:type 1 host 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:type 2 chassis 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:type 3 rack 2026-03-10T13:00:21.330 INFO:teuthology.orchestra.run.vm04.stderr:type 4 row 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 5 pdu 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 6 pod 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 7 room 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 8 datacenter 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 9 zone 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 10 region 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:type 11 root 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:# buckets 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:host vm04 { 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -3 # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -4 class hdd # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.03897 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item osd.0 weight 0.01949 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item osd.1 weight 0.01949 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:host vm06 { 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -5 # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -6 class hdd # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.03897 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item osd.2 weight 0.01949 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item osd.3 weight 0.01949 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:root default { 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -1 # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id -2 class hdd # do not change unnecessarily 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: # weight 0.07794 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: alg straw2 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: hash 0 # rjenkins1 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item vm04 weight 0.03897 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: item vm06 weight 0.03897 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:# rules 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:rule replicated_rule { 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: id 0 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: type replicated 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: step take default 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: step choose firstn 0 type osd 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: step emit 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:} 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:# end crush map' 2026-03-10T13:00:21.331 INFO:teuthology.orchestra.run.vm04.stderr:+ grep -q vm09 2026-03-10T13:00:21.390 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T13:00:21.392 INFO:tasks.cephadm:Teardown begin 2026-03-10T13:00:21.392 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:21.415 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:21 vm04 ceph-mon[50620]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:21.415 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:21 vm04 ceph-mon[50620]: from='client.? 192.168.123.104:0/963269417' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-10T13:00:21.421 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:21.447 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:21.472 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T13:00:21.472 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 -- ceph mgr module disable cephadm 2026-03-10T13:00:21.629 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/mon.a/config 2026-03-10T13:00:21.649 INFO:teuthology.orchestra.run.vm04.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-10T13:00:21.670 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-10T13:00:21.670 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T13:00:21.671 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T13:00:21.686 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T13:00:21.701 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T13:00:21.716 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T13:00:21.716 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-10T13:00:21.716 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a 2026-03-10T13:00:21.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:21 vm06 ceph-mon[56046]: pgmap v120: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-10T13:00:21.859 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:21 vm06 ceph-mon[56046]: from='client.? 192.168.123.104:0/963269417' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-10T13:00:22.079 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:21 vm04 systemd[1]: Stopping Ceph mon.a for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:22.079 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a[50616]: 2026-03-10T13:00:21.821+0000 7fd66d285640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:22.079 INFO:journalctl@ceph.mon.a.vm04.stdout:Mar 10 13:00:21 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-a[50616]: 2026-03-10T13:00:21.821+0000 7fd66d285640 -1 mon.a@0(leader) e4 *** Got Signal Terminated *** 2026-03-10T13:00:22.168 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.a.service' 2026-03-10T13:00:22.201 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:22.201 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-10T13:00:22.202 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-10T13:00:22.202 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.b 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 systemd[1]: Stopping Ceph mon.b for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-b[56042]: 2026-03-10T13:00:22.313+0000 7f9aa42b1640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-b[56042]: 2026-03-10T13:00:22.313+0000 7f9aa42b1640 -1 mon.b@1(peon) e4 *** Got Signal Terminated *** 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 podman[69167]: 2026-03-10 13:00:22.446775273 +0000 UTC m=+0.148856452 container died dc17a80cdb2fad9062abe75507574f2e4f686fb075e613217dec467058fac68c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.build-date=20260223) 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 podman[69167]: 2026-03-10 13:00:22.463648149 +0000 UTC m=+0.165729328 container remove dc17a80cdb2fad9062abe75507574f2e4f686fb075e613217dec467058fac68c (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-b, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_REF=squid, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T13:00:22.516 INFO:journalctl@ceph.mon.b.vm06.stdout:Mar 10 13:00:22 vm06 bash[69167]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-mon-b 2026-03-10T13:00:22.527 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.b.service' 2026-03-10T13:00:22.571 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:22.571 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-10T13:00:22.571 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-10T13:00:22.571 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.c 2026-03-10T13:00:22.608 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mon.c.service' 2026-03-10T13:00:22.684 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:22.684 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-10T13:00:22.684 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-10T13:00:22.684 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a 2026-03-10T13:00:22.910 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.a.service' 2026-03-10T13:00:22.944 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:22.944 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-10T13:00:22.944 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-10T13:00:22.944 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.b 2026-03-10T13:00:23.080 INFO:journalctl@ceph.mgr.b.vm06.stdout:Mar 10 13:00:22 vm06 systemd[1]: Stopping Ceph mgr.b for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:23.177 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@mgr.b.service' 2026-03-10T13:00:23.215 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:23.215 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-10T13:00:23.215 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-10T13:00:23.215 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.0 2026-03-10T13:00:23.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:23 vm04 systemd[1]: Stopping Ceph osd.0 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:23.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T13:00:23.320+0000 7efd4f68a640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:23.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T13:00:23.320+0000 7efd4f68a640 -1 osd.0 47 *** Got signal Terminated *** 2026-03-10T13:00:23.626 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:23 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0[62589]: 2026-03-10T13:00:23.320+0000 7efd4f68a640 -1 osd.0 47 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75817]: 2026-03-10 13:00:28.349091475 +0000 UTC m=+5.043231021 container died ac5275d8bf64a4a1d1342e6a5561591b7677fd97b5758ca38229fd6b9f56f373 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75817]: 2026-03-10 13:00:28.380536377 +0000 UTC m=+5.074675923 container remove ac5275d8bf64a4a1d1342e6a5561591b7677fd97b5758ca38229fd6b9f56f373 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0) 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 bash[75817]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75883]: 2026-03-10 13:00:28.545812808 +0000 UTC m=+0.023127989 container create fa5e05116c080ac788eab2140520641044d5449dcef5c4d840fd09117f258ace (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, ceph=True, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75883]: 2026-03-10 13:00:28.593781052 +0000 UTC m=+0.071096233 container init fa5e05116c080ac788eab2140520641044d5449dcef5c4d840fd09117f258ace (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0-deactivate, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75883]: 2026-03-10 13:00:28.596814341 +0000 UTC m=+0.074129522 container start fa5e05116c080ac788eab2140520641044d5449dcef5c4d840fd09117f258ace (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0-deactivate, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_REF=squid, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, OSD_FLAVOR=default) 2026-03-10T13:00:28.603 INFO:journalctl@ceph.osd.0.vm04.stdout:Mar 10 13:00:28 vm04 podman[75883]: 2026-03-10 13:00:28.600663569 +0000 UTC m=+0.077978750 container attach fa5e05116c080ac788eab2140520641044d5449dcef5c4d840fd09117f258ace (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-0-deactivate, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, ceph=True, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T13:00:28.779 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.0.service' 2026-03-10T13:00:28.815 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:28.815 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-10T13:00:28.815 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-10T13:00:28.815 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.1 2026-03-10T13:00:29.376 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:28 vm04 systemd[1]: Stopping Ceph osd.1 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:29.376 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:28 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T13:00:28.959+0000 7f610de12640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:29.376 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:28 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T13:00:28.959+0000 7f610de12640 -1 osd.1 47 *** Got signal Terminated *** 2026-03-10T13:00:29.376 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:28 vm04 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1[67788]: 2026-03-10T13:00:28.959+0000 7f610de12640 -1 osd.1 47 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T13:00:34.234 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:33 vm04 podman[75978]: 2026-03-10 13:00:33.97850069 +0000 UTC m=+5.031803179 container died 1523d955674c11ea99053fb649bacf9d513702ea5700656973d80f0d51df2757 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, ceph=True, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default) 2026-03-10T13:00:34.234 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 podman[75978]: 2026-03-10 13:00:34.010434479 +0000 UTC m=+5.063736968 container remove 1523d955674c11ea99053fb649bacf9d513702ea5700656973d80f0d51df2757 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T13:00:34.235 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 bash[75978]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1 2026-03-10T13:00:34.235 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 podman[76045]: 2026-03-10 13:00:34.13929613 +0000 UTC m=+0.015862453 container create 110df9dfe6302d3a1d35984d51f477f73bb742783b82231637f7c07db10cb28b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1-deactivate, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2) 2026-03-10T13:00:34.235 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 podman[76045]: 2026-03-10 13:00:34.175193541 +0000 UTC m=+0.051759874 container init 110df9dfe6302d3a1d35984d51f477f73bb742783b82231637f7c07db10cb28b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, OSD_FLAVOR=default, io.buildah.version=1.41.3, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS) 2026-03-10T13:00:34.235 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 podman[76045]: 2026-03-10 13:00:34.181827544 +0000 UTC m=+0.058393856 container start 110df9dfe6302d3a1d35984d51f477f73bb742783b82231637f7c07db10cb28b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.build-date=20260223, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, OSD_FLAVOR=default) 2026-03-10T13:00:34.235 INFO:journalctl@ceph.osd.1.vm04.stdout:Mar 10 13:00:34 vm04 podman[76045]: 2026-03-10 13:00:34.182789906 +0000 UTC m=+0.059356229 container attach 110df9dfe6302d3a1d35984d51f477f73bb742783b82231637f7c07db10cb28b (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-1-deactivate, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, CEPH_REF=squid) 2026-03-10T13:00:34.358 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.1.service' 2026-03-10T13:00:34.389 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:34.389 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-10T13:00:34.389 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-10T13:00:34.389 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.2 2026-03-10T13:00:34.860 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:34 vm06 systemd[1]: Stopping Ceph osd.2 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:34.860 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:34 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T13:00:34.484+0000 7f6968ffa640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:34.860 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:34 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T13:00:34.484+0000 7f6968ffa640 -1 osd.2 47 *** Got signal Terminated *** 2026-03-10T13:00:34.860 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:34 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2[59686]: 2026-03-10T13:00:34.484+0000 7f6968ffa640 -1 osd.2 47 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69373]: 2026-03-10 13:00:39.516027109 +0000 UTC m=+5.042349188 container died 7fa95984b9fceae797a586ede7e69084026b5897e51811b280f46809803484ba (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69373]: 2026-03-10 13:00:39.533984333 +0000 UTC m=+5.060306412 container remove 7fa95984b9fceae797a586ede7e69084026b5897e51811b280f46809803484ba (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, ceph=True, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 bash[69373]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69441]: 2026-03-10 13:00:39.684066169 +0000 UTC m=+0.018504031 container create 59293c20e1448dc347b98afec6572293b7dee3b408bccd5386999b83fd4e1008 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2-deactivate, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_REF=squid, ceph=True) 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69441]: 2026-03-10 13:00:39.725279177 +0000 UTC m=+0.059717039 container init 59293c20e1448dc347b98afec6572293b7dee3b408bccd5386999b83fd4e1008 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2-deactivate, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, ceph=True) 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69441]: 2026-03-10 13:00:39.728271368 +0000 UTC m=+0.062709230 container start 59293c20e1448dc347b98afec6572293b7dee3b408bccd5386999b83fd4e1008 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True) 2026-03-10T13:00:39.775 INFO:journalctl@ceph.osd.2.vm06.stdout:Mar 10 13:00:39 vm06 podman[69441]: 2026-03-10 13:00:39.729075624 +0000 UTC m=+0.063513486 container attach 59293c20e1448dc347b98afec6572293b7dee3b408bccd5386999b83fd4e1008 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-2-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-10T13:00:39.920 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.2.service' 2026-03-10T13:00:39.961 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:39.961 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-10T13:00:39.961 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-10T13:00:39.961 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.3 2026-03-10T13:00:40.360 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:40 vm06 systemd[1]: Stopping Ceph osd.3 for 74d0e8a6-1c80-11f1-9542-4d7926099318... 2026-03-10T13:00:40.360 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:40 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T13:00:40.102+0000 7ffb14724640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T13:00:40.360 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:40 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T13:00:40.102+0000 7ffb14724640 -1 osd.3 47 *** Got signal Terminated *** 2026-03-10T13:00:40.360 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:40 vm06 ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3[64900]: 2026-03-10T13:00:40.102+0000 7ffb14724640 -1 osd.3 47 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69538]: 2026-03-10 13:00:45.12630001 +0000 UTC m=+5.035986459 container died 0c54f38d691ad07d5c5aef8866c4e807e559248fae2a49693db7e26edc08e4a3 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69538]: 2026-03-10 13:00:45.157120895 +0000 UTC m=+5.066807334 container remove 0c54f38d691ad07d5c5aef8866c4e807e559248fae2a49693db7e26edc08e4a3 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3, org.label-schema.license=GPLv2, ceph=True, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 bash[69538]: ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69606]: 2026-03-10 13:00:45.299917933 +0000 UTC m=+0.016958496 container create fead5b73d2d1c11ab2469ba08e9b40cfe86329ec0d6f52bc74bd0c2c979bd7ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.license=GPLv2, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69606]: 2026-03-10 13:00:45.338606396 +0000 UTC m=+0.055646949 container init fead5b73d2d1c11ab2469ba08e9b40cfe86329ec0d6f52bc74bd0c2c979bd7ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3-deactivate, org.label-schema.license=GPLv2, ceph=True, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69606]: 2026-03-10 13:00:45.341921181 +0000 UTC m=+0.058961734 container start fead5b73d2d1c11ab2469ba08e9b40cfe86329ec0d6f52bc74bd0c2c979bd7ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3-deactivate, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69606]: 2026-03-10 13:00:45.343909182 +0000 UTC m=+0.060949745 container attach fead5b73d2d1c11ab2469ba08e9b40cfe86329ec0d6f52bc74bd0c2c979bd7ca (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-74d0e8a6-1c80-11f1-9542-4d7926099318-osd-3-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223) 2026-03-10T13:00:45.394 INFO:journalctl@ceph.osd.3.vm06.stdout:Mar 10 13:00:45 vm06 podman[69606]: 2026-03-10 13:00:45.293384202 +0000 UTC m=+0.010424775 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T13:00:45.502 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.3.service' 2026-03-10T13:00:45.539 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:45.539 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-10T13:00:45.539 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-10T13:00:45.539 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.4 2026-03-10T13:00:45.575 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.4.service' 2026-03-10T13:00:45.643 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:45.643 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-10T13:00:45.643 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-10T13:00:45.643 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.5 2026-03-10T13:00:45.717 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-74d0e8a6-1c80-11f1-9542-4d7926099318@osd.5.service' 2026-03-10T13:00:45.788 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T13:00:45.788 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-10T13:00:45.788 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force --keep-logs 2026-03-10T13:00:45.943 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:47.109 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force --keep-logs 2026-03-10T13:00:47.242 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:48.440 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force --keep-logs 2026-03-10T13:00:48.574 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:48.962 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:48.996 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:49.022 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T13:00:49.050 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T13:00:49.050 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm04/crash 2026-03-10T13:00:49.050 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash -- . 2026-03-10T13:00:49.080 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash: Cannot open: No such file or directory 2026-03-10T13:00:49.080 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-10T13:00:49.081 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm06/crash 2026-03-10T13:00:49.081 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash -- . 2026-03-10T13:00:49.109 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash: Cannot open: No such file or directory 2026-03-10T13:00:49.109 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-10T13:00:49.110 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm09/crash 2026-03-10T13:00:49.110 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash -- . 2026-03-10T13:00:49.136 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/crash: Cannot open: No such file or directory 2026-03-10T13:00:49.136 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-10T13:00:49.137 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T13:00:49.137 DEBUG:teuthology.orchestra.run.vm04:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v 'mons down' | egrep -v 'mon down' | egrep -v 'out of quorum' | egrep -v CEPHADM_STRAY_HOST | egrep -v CEPHADM_STRAY_DAEMON | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-10T13:00:49.175 INFO:tasks.cephadm:Compressing logs... 2026-03-10T13:00:49.175 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:00:49.217 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:00:49.219 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:00:49.241 INFO:teuthology.orchestra.run.vm04.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T13:00:49.241 INFO:teuthology.orchestra.run.vm04.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T13:00:49.242 INFO:teuthology.orchestra.run.vm06.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T13:00:49.243 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T13:00:49.243 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.a.log 2026-03-10T13:00:49.243 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log 2026-03-10T13:00:49.243 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log 2026-03-10T13:00:49.244 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: 90.0% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T13:00:49.244 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.b.log 2026-03-10T13:00:49.244 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.a.log: 90.6% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T13:00:49.245 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.a.log 2026-03-10T13:00:49.245 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T13:00:49.245 INFO:teuthology.orchestra.run.vm09.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T13:00:49.245 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log: 85.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log.gz 2026-03-10T13:00:49.245 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log 2026-03-10T13:00:49.246 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log 2026-03-10T13:00:49.246 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.c.log 2026-03-10T13:00:49.247 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log: 89.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T13:00:49.247 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log 2026-03-10T13:00:49.247 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log: 95.1% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log.gz 2026-03-10T13:00:49.247 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log 2026-03-10T13:00:49.248 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log 2026-03-10T13:00:49.248 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.b.log: 82.8% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log.gz 2026-03-10T13:00:49.248 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log 2026-03-10T13:00:49.249 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log: 90.7% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log.gz 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.b.log 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.c.log: 95.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log.gz 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log: 85.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log.gz 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log: 80.4% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log.gz 2026-03-10T13:00:49.250 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.2.log 2026-03-10T13:00:49.252 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.b.log: 90.3% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.b.log.gz 2026-03-10T13:00:49.252 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.3.log 2026-03-10T13:00:49.254 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log 2026-03-10T13:00:49.255 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log: 90.5% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log.gz 2026-03-10T13:00:49.255 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log 2026-03-10T13:00:49.255 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.4.log 2026-03-10T13:00:49.256 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log: 83.7% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.log.gz 2026-03-10T13:00:49.256 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.5.log 2026-03-10T13:00:49.256 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log: 90.4% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.audit.log.gz 2026-03-10T13:00:49.256 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log 2026-03-10T13:00:49.257 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph.cephadm.log.gz 2026-03-10T13:00:49.267 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.0.log 2026-03-10T13:00:49.267 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.4.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.5.log: 93.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.4.log.gz 2026-03-10T13:00:49.270 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log: 95.1% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-volume.log.gz 2026-03-10T13:00:49.270 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.1.log 2026-03-10T13:00:49.272 INFO:teuthology.orchestra.run.vm09.stderr: 93.1% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.5.log.gz 2026-03-10T13:00:49.277 INFO:teuthology.orchestra.run.vm09.stderr: 92.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.c.log.gz 2026-03-10T13:00:49.278 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-10T13:00:49.278 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.044s 2026-03-10T13:00:49.278 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.055s 2026-03-10T13:00:49.278 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.017s 2026-03-10T13:00:49.280 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.2.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.3.log: 92.2% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.b.log.gz 2026-03-10T13:00:49.290 INFO:teuthology.orchestra.run.vm06.stderr: 93.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.2.log.gz 2026-03-10T13:00:49.293 INFO:teuthology.orchestra.run.vm06.stderr: 93.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.3.log.gz 2026-03-10T13:00:49.295 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-10T13:00:49.295 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.062s 2026-03-10T13:00:49.295 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.083s 2026-03-10T13:00:49.295 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.027s 2026-03-10T13:00:49.326 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.0.log: /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.1.log: 92.7% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.0.log.gz 2026-03-10T13:00:49.326 INFO:teuthology.orchestra.run.vm04.stderr: 93.0% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-osd.1.log.gz 2026-03-10T13:00:49.326 INFO:teuthology.orchestra.run.vm04.stderr: 89.9% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mgr.a.log.gz 2026-03-10T13:00:49.355 INFO:teuthology.orchestra.run.vm04.stderr: 91.4% -- replaced with /var/log/ceph/74d0e8a6-1c80-11f1-9542-4d7926099318/ceph-mon.a.log.gz 2026-03-10T13:00:49.356 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-10T13:00:49.356 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.125s 2026-03-10T13:00:49.357 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.195s 2026-03-10T13:00:49.357 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.019s 2026-03-10T13:00:49.357 INFO:tasks.cephadm:Archiving logs... 2026-03-10T13:00:49.357 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm04/log 2026-03-10T13:00:49.357 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T13:00:49.436 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm06/log 2026-03-10T13:00:49.436 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T13:00:49.472 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm09/log 2026-03-10T13:00:49.472 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T13:00:49.500 INFO:tasks.cephadm:Removing cluster... 2026-03-10T13:00:49.500 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force 2026-03-10T13:00:49.629 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:49.855 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force 2026-03-10T13:00:49.985 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:50.197 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 74d0e8a6-1c80-11f1-9542-4d7926099318 --force 2026-03-10T13:00:50.328 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 74d0e8a6-1c80-11f1-9542-4d7926099318 2026-03-10T13:00:50.550 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T13:00:50.550 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T13:00:50.566 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T13:00:50.585 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T13:00:50.604 INFO:tasks.cephadm:Teardown complete 2026-03-10T13:00:50.604 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-10T13:00:50.607 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-10T13:00:50.607 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T13:00:50.609 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T13:00:50.627 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-10T13:00:50.677 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T13:00:50.678 DEBUG:teuthology.orchestra.run.vm04:> 2026-03-10T13:00:50.678 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T13:00:50.678 DEBUG:teuthology.orchestra.run.vm04:> sudo yum -y remove $d || true 2026-03-10T13:00:50.678 DEBUG:teuthology.orchestra.run.vm04:> done 2026-03-10T13:00:50.684 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T13:00:50.684 DEBUG:teuthology.orchestra.run.vm06:> 2026-03-10T13:00:50.684 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T13:00:50.684 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-03-10T13:00:50.684 DEBUG:teuthology.orchestra.run.vm06:> done 2026-03-10T13:00:50.691 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-10T13:00:50.691 DEBUG:teuthology.orchestra.run.vm09:> 2026-03-10T13:00:50.691 DEBUG:teuthology.orchestra.run.vm09:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-10T13:00:50.691 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y remove $d || true 2026-03-10T13:00:50.691 DEBUG:teuthology.orchestra.run.vm09:> done 2026-03-10T13:00:50.866 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 39 M 2026-03-10T13:00:50.867 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:00:50.869 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:00:50.869 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 39 M 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:00:50.882 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:00:50.883 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:00:50.885 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:00:50.885 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:00:50.892 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Remove 2 Packages 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 39 M 2026-03-10T13:00:50.893 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:00:50.895 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:00:50.896 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:00:50.899 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:00:50.899 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:00:50.909 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:00:50.909 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:00:50.915 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:00:50.937 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.940 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:50.942 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:00:50.944 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.954 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.961 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:50.964 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.966 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.966 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:50.967 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-10T13:00:50.967 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.967 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-10T13:00:50.967 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:50.969 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:50.970 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.974 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.980 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:50.990 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:50.995 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.042 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.043 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:51.072 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.072 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:51.074 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.074 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.105 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.128 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.130 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:00:51.343 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Remove 4 Packages 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 212 M 2026-03-10T13:00:51.344 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:00:51.346 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Remove 4 Packages 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 212 M 2026-03-10T13:00:51.347 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:00:51.348 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:00:51.348 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:00:51.350 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:00:51.350 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:00:51.367 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Remove 4 Packages 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 212 M 2026-03-10T13:00:51.368 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:00:51.371 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:00:51.372 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:00:51.374 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:00:51.374 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:00:51.375 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:00:51.376 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:00:51.400 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:00:51.400 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:00:51.441 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:00:51.442 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:00:51.449 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.449 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.451 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T13:00:51.451 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T13:00:51.455 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T13:00:51.455 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T13:00:51.466 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:00:51.471 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.471 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.473 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.476 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-10T13:00:51.479 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-10T13:00:51.495 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.542 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.542 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.542 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T13:00:51.542 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T13:00:51.543 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.543 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.543 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T13:00:51.543 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T13:00:51.570 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-10T13:00:51.570 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-10T13:00:51.570 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-10T13:00:51.570 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-10T13:00:51.601 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.602 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.603 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.627 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:00:51.824 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Remove 8 Packages 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 28 M 2026-03-10T13:00:51.825 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:00:51.830 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:00:51.830 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:00:51.841 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Remove 8 Packages 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 28 M 2026-03-10T13:00:51.842 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:00:51.845 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:00:51.845 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:00:51.854 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:00:51.855 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:00:51.862 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-03-10T13:00:51.863 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.864 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-03-10T13:00:51.864 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:00:51.866 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:00:51.866 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:00:51.868 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:00:51.868 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:00:51.891 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:00:51.891 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:00:51.902 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:00:51.908 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:51.911 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T13:00:51.912 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:00:51.913 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T13:00:51.916 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T13:00:51.917 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:51.919 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T13:00:51.921 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T13:00:51.921 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T13:00:51.922 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T13:00:51.925 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T13:00:51.928 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T13:00:51.930 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T13:00:51.934 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T13:00:51.940 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.941 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.943 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-10T13:00:51.945 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-10T13:00:51.948 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.948 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T13:00:51.950 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.951 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.951 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-10T13:00:51.953 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-10T13:00:51.960 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T13:00:51.966 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:51.967 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:51.975 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.975 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:51.976 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-10T13:00:51.976 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-10T13:00:51.976 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-10T13:00:51.976 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:51.976 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T13:00:51.982 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:51.984 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:51.984 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-10T13:00:52.005 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.007 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T13:00:52.056 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T13:00:52.084 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.112 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-10T13:00:52.116 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-10T13:00:52.205 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T13:00:52.205 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: zip-3.0-35.el9.x86_64 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.206 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:52.258 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.259 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:00:52.384 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T13:00:52.390 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T13:00:52.391 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout:=========================================================================================== 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout:Remove 102 Packages 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 613 M 2026-03-10T13:00:52.392 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:00:52.422 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:00:52.424 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:00:52.424 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:00:52.428 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T13:00:52.429 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout:=========================================================================================== 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout:Remove 102 Packages 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 613 M 2026-03-10T13:00:52.430 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:00:52.459 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:00:52.459 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:00:52.480 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-10T13:00:52.486 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-10T13:00:52.487 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout:=========================================================================================== 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout:Remove 102 Packages 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 613 M 2026-03-10T13:00:52.488 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:00:52.518 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:00:52.518 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:00:52.534 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:00:52.535 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:00:52.567 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:00:52.568 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:00:52.632 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:00:52.632 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:00:52.693 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:00:52.693 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.701 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.722 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.723 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:00:52.723 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.723 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.731 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.741 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.748 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.749 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.762 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.765 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T13:00:52.766 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.787 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T13:00:52.788 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.816 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:00:52.817 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.826 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-10T13:00:52.826 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.836 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T13:00:52.841 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T13:00:52.841 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-10T13:00:52.844 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:52.845 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.851 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.854 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.860 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:52.860 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T13:00:52.862 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T13:00:52.865 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T13:00:52.865 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.866 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T13:00:52.875 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T13:00:52.877 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.879 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T13:00:52.885 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T13:00:52.885 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-10T13:00:52.885 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.888 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T13:00:52.898 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T13:00:52.902 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T13:00:52.903 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.908 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.918 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.924 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.924 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.925 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T13:00:52.925 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T13:00:52.925 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T13:00:52.925 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.933 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.938 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.938 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.938 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T13:00:52.938 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:52.942 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:52.945 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-10T13:00:52.946 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.955 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-10T13:00:52.956 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.959 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T13:00:52.959 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.959 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:52.959 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T13:00:52.959 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:52.960 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-10T13:00:52.960 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.964 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T13:00:52.968 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.969 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T13:00:52.973 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:52.977 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T13:00:52.979 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:52.980 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-10T13:00:52.982 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T13:00:52.985 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-10T13:00:52.987 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T13:00:52.990 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T13:00:52.992 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T13:00:52.993 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-10T13:00:52.996 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T13:00:52.998 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-10T13:00:53.001 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T13:00:53.006 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T13:00:53.012 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T13:00:53.015 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T13:00:53.018 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:53.019 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.019 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-10T13:00:53.019 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-10T13:00:53.019 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-10T13:00:53.019 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.022 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T13:00:53.024 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:53.032 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T13:00:53.034 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:53.039 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T13:00:53.043 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T13:00:53.049 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T13:00:53.050 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:53.050 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.050 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-10T13:00:53.050 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.052 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T13:00:53.059 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:53.061 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T13:00:53.070 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-10T13:00:53.070 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T13:00:53.071 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T13:00:53.071 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.073 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-10T13:00:53.079 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T13:00:53.079 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-10T13:00:53.079 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.082 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T13:00:53.085 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-10T13:00:53.092 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T13:00:53.094 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-10T13:00:53.104 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T13:00:53.104 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.107 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-10T13:00:53.114 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.115 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-10T13:00:53.126 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-10T13:00:53.133 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-10T13:00:53.163 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-10T13:00:53.170 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-10T13:00:53.173 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-10T13:00:53.176 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T13:00:53.182 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-10T13:00:53.193 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T13:00:53.194 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-10T13:00:53.194 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.203 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-10T13:00:53.208 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.208 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T13:00:53.208 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:53.209 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.212 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T13:00:53.229 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T13:00:53.240 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.246 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.246 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T13:00:53.246 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:53.247 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.256 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T13:00:53.263 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T13:00:53.266 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T13:00:53.269 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T13:00:53.274 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:53.290 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T13:00:53.291 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.297 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T13:00:53.300 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T13:00:53.302 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-10T13:00:53.302 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T13:00:53.303 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.308 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T13:00:53.310 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T13:00:53.318 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.322 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:53.324 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.324 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T13:00:53.327 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T13:00:53.332 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T13:00:53.337 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T13:00:53.337 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.338 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.338 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-10T13:00:53.338 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.339 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.342 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T13:00:53.345 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T13:00:53.348 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T13:00:53.351 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T13:00:53.355 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T13:00:53.359 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T13:00:53.367 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-10T13:00:53.386 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T13:00:53.391 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T13:00:53.415 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-10T13:00:53.430 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T13:00:53.441 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T13:00:53.449 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-10T13:00:53.451 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T13:00:53.452 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-10T13:00:53.453 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T13:00:53.455 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-10T13:00:53.456 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T13:00:53.461 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T13:00:53.463 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T13:00:53.464 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T13:00:53.465 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T13:00:53.471 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T13:00:53.473 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-10T13:00:53.474 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.475 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.476 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T13:00:53.479 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T13:00:53.486 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-10T13:00:53.489 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.489 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.489 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T13:00:53.490 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:53.490 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.493 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-10T13:00:53.495 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-10T13:00:53.497 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-10T13:00:53.500 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-10T13:00:53.500 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.501 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.501 INFO:teuthology.orchestra.run.vm04.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.501 INFO:teuthology.orchestra.run.vm04.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T13:00:53.501 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:53.501 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.503 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T13:00:53.504 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-10T13:00:53.505 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T13:00:53.508 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T13:00:53.508 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-10T13:00:53.510 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.511 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T13:00:53.512 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T13:00:53.513 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-10T13:00:53.513 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T13:00:53.514 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T13:00:53.516 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T13:00:53.517 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T13:00:53.519 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T13:00:53.520 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T13:00:53.522 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T13:00:53.522 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T13:00:53.525 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T13:00:53.528 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T13:00:53.530 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T13:00:53.531 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T13:00:53.535 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T13:00:53.537 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T13:00:53.541 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T13:00:53.541 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T13:00:53.544 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T13:00:53.545 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T13:00:53.547 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T13:00:53.550 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T13:00:53.550 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T13:00:53.553 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T13:00:53.554 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T13:00:53.559 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T13:00:53.559 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T13:00:53.562 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-10T13:00:53.564 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T13:00:53.564 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T13:00:53.570 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T13:00:53.571 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T13:00:53.574 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T13:00:53.575 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-10T13:00:53.575 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T13:00:53.577 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-10T13:00:53.578 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T13:00:53.580 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T13:00:53.581 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T13:00:53.583 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-10T13:00:53.584 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T13:00:53.585 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-10T13:00:53.586 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T13:00:53.588 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T13:00:53.589 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-10T13:00:53.590 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T13:00:53.592 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-10T13:00:53.592 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T13:00:53.595 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T13:00:53.599 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T13:00:53.609 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T13:00:53.613 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T13:00:53.613 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.614 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-10T13:00:53.614 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-10T13:00:53.614 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.614 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T13:00:53.614 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.670 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T13:00:53.671 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T13:00:53.685 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-10T13:00:53.686 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T13:00:53.687 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-10T13:00:53.689 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T13:00:53.694 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T13:00:53.701 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-10T13:00:53.703 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T13:00:53.703 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T13:00:53.705 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-10T13:00:53.706 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T13:00:53.707 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-10T13:00:53.708 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T13:00:53.709 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T13:00:53.709 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-10T13:00:53.712 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-10T13:00:53.712 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T13:00:53.714 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T13:00:53.715 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-10T13:00:53.717 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T13:00:53.718 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-10T13:00:53.726 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-10T13:00:53.731 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-10T13:00:53.732 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.732 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T13:00:53.732 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:53.733 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-10T13:00:53.735 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-10T13:00:53.738 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.738 INFO:teuthology.orchestra.run.vm04.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T13:00:53.738 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:53.738 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-10T13:00:53.740 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.744 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-10T13:00:53.745 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.748 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-10T13:00:53.754 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-10T13:00:53.758 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-10T13:00:53.764 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-10T13:00:53.767 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-10T13:00:53.768 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.768 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.770 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-10T13:00:53.773 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-10T13:00:53.774 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.774 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.779 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-10T13:00:53.780 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.782 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-10T13:00:53.785 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T13:00:53.787 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.787 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-10T13:00:53.788 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T13:00:53.789 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T13:00:53.789 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:53.791 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T13:00:53.794 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T13:00:53.795 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-10T13:00:53.796 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T13:00:53.796 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:53.800 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-10T13:00:53.804 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-10T13:00:53.806 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-10T13:00:53.808 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-10T13:00:53.813 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-10T13:00:53.817 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-10T13:00:53.838 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.838 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-10T13:00:53.838 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:53.846 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.875 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-10T13:00:53.875 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.888 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-10T13:00:53.893 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-10T13:00:53.897 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-10T13:00:53.899 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-10T13:00:53.899 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /sys 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /proc 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /mnt 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /var/tmp 2026-03-10T13:00:59.497 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /home 2026-03-10T13:00:59.498 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /root 2026-03-10T13:00:59.498 INFO:teuthology.orchestra.run.vm04.stdout:skipping the directory /tmp 2026-03-10T13:00:59.498 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:59.506 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T13:00:59.524 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.524 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.532 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.535 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T13:00:59.538 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T13:00:59.540 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T13:00:59.542 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T13:00:59.542 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /sys 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /proc 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /mnt 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /var/tmp 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /home 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /root 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /tmp 2026-03-10T13:00:59.553 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:59.556 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.558 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T13:00:59.561 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T13:00:59.562 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T13:00:59.563 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T13:00:59.566 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T13:00:59.572 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T13:00:59.580 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T13:00:59.585 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.585 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.612 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T13:00:59.612 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.612 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.616 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T13:00:59.619 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T13:00:59.621 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T13:00:59.623 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T13:00:59.623 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.645 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.647 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T13:00:59.650 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T13:00:59.653 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T13:00:59.657 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T13:00:59.664 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T13:00:59.673 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T13:00:59.677 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T13:00:59.677 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T13:00:59.716 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T13:00:59.717 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T13:00:59.718 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-03-10T13:00:59.755 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:00:59.760 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T13:00:59.776 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T13:00:59.777 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T13:00:59.778 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T13:00:59.779 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T13:00:59.780 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T13:00:59.783 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-10T13:00:59.786 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-10T13:00:59.788 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-10T13:00:59.792 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-10T13:00:59.793 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-10T13:00:59.794 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.805 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T13:00:59.806 INFO:teuthology.orchestra.run.vm04.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:00:59.807 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:00:59.810 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-10T13:00:59.812 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-10T13:00:59.822 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-10T13:00:59.825 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-10T13:00:59.832 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-10T13:00:59.843 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-10T13:00:59.848 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-10T13:00:59.848 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.858 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T13:00:59.859 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:00:59.860 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-10T13:00:59.966 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-10T13:00:59.967 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-10T13:00:59.968 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 775 k 2026-03-10T13:01:00.036 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:01:00.038 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:01:00.038 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:01:00.039 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:01:00.039 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.043 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-10T13:01:00.044 INFO:teuthology.orchestra.run.vm06.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.045 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:00.056 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:01:00.056 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.083 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:Remove 1 Package 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 775 k 2026-03-10T13:01:00.084 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:01:00.086 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:01:00.086 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:01:00.088 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:01:00.088 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:01:00.105 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:01:00.106 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.167 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:00.217 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:00.219 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.261 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:00.261 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:00.261 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:01:00.261 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:00.261 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 775 k 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:00.262 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:01:00.263 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:01:00.263 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:01:00.264 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:01:00.265 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:01:00.281 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:01:00.281 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.390 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.408 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T13:01:00.409 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:00.412 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:00.412 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:00.413 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:00.441 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-10T13:01:00.441 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.441 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:00.441 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-10T13:01:00.441 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:00.442 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:00.466 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T13:01:00.466 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:00.470 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:00.470 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:00.470 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:00.579 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr 2026-03-10T13:01:00.579 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:00.582 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:00.583 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:00.583 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:00.655 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr 2026-03-10T13:01:00.655 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:00.658 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:00.659 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:00.659 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:00.664 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-immutable-object-cache 2026-03-10T13:01:00.664 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:00.667 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:00.668 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:00.668 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:00.758 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T13:01:00.758 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:00.761 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:00.762 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:00.762 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:00.843 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T13:01:00.843 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:00.847 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:00.847 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:00.847 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:00.852 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr 2026-03-10T13:01:00.852 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:00.856 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:00.856 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:00.856 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:00.945 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T13:01:00.945 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:00.948 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:00.949 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:00.949 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:01.032 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T13:01:01.032 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:01.035 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:01.036 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:01.036 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:01.056 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-dashboard 2026-03-10T13:01:01.056 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:01.060 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:01.060 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:01.060 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:01.144 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-rook 2026-03-10T13:01:01.144 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:01.148 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:01.148 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:01.149 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:01.229 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-rook 2026-03-10T13:01:01.229 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:01.233 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:01.234 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:01.234 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:01.254 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-10T13:01:01.255 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:01.258 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:01.258 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:01.259 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:01.335 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T13:01:01.335 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:01.338 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:01.339 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:01.339 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:01.411 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T13:01:01.411 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:01.414 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:01.415 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:01.415 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:01.440 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-rook 2026-03-10T13:01:01.441 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:01.444 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:01.444 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:01.444 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:01:01.512 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:01.513 INFO:teuthology.orchestra.run.vm04.stdout:Remove 1 Package 2026-03-10T13:01:01.513 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:01.513 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.6 M 2026-03-10T13:01:01.513 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:01:01.514 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:01:01.515 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:01:01.524 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:01:01.525 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:01:01.550 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:01:01.565 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.610 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:Remove 1 Package 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 3.6 M 2026-03-10T13:01:01.611 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:01:01.613 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:01:01.613 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:01:01.623 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:01:01.624 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:01:01.629 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-mgr-cephadm 2026-03-10T13:01:01.629 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:01.632 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:01.633 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:01.633 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:01.636 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.651 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:01:01.666 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.686 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.686 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:01.687 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:01:01.687 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:01.687 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:01.687 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:01.741 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:01.781 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Remove 1 Package 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.6 M 2026-03-10T13:01:01.839 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:01:01.841 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:01:01.841 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:01:01.851 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:01:01.852 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:01:01.877 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:01:01.892 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.899 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: ceph-volume 2026-03-10T13:01:01.899 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:01.902 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:01.903 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:01.903 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:01.970 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:01.984 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: ceph-volume 2026-03-10T13:01:01.984 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:01.987 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:01.987 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:01.987 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.021 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:02.094 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Remove 2 Packages 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 610 k 2026-03-10T13:01:02.095 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:01:02.097 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:01:02.097 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:01:02.108 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:01:02.108 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:01:02.138 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:01:02.140 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.154 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.189 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Remove 2 Packages 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 610 k 2026-03-10T13:01:02.190 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:01:02.192 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:01:02.192 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:01:02.203 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:01:02.203 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:01:02.204 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: ceph-volume 2026-03-10T13:01:02.204 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:02.207 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:02.208 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:02.208 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:02.219 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.219 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.228 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:01:02.230 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.244 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.260 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:02.312 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.312 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.376 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 610 k 2026-03-10T13:01:02.404 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:01:02.406 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:01:02.406 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:01:02.417 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:01:02.418 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:01:02.444 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:01:02.447 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.453 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Remove 3 Packages 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 3.7 M 2026-03-10T13:01:02.454 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:01:02.456 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:01:02.456 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:01:02.460 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.473 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:01:02.473 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:01:02.505 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:01:02.507 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.508 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.508 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.533 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.533 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:01:02.576 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Remove 3 Packages 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 3.7 M 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.577 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:01:02.579 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:01:02.579 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.582 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:02.596 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:01:02.597 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.617 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.618 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:02.632 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:01:02.634 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.636 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.636 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.702 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.702 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.702 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.747 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.747 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.747 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:01:02.747 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.747 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.748 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.748 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:02.748 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-10T13:01:02.785 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 3.7 M 2026-03-10T13:01:02.786 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:01:02.787 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:01:02.787 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:01:02.795 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: libcephfs-devel 2026-03-10T13:01:02.795 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:02.799 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:02.800 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:02.800 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:02.806 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:01:02.806 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:01:02.842 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:01:02.844 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.846 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.846 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.911 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.912 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-10T13:01:02.912 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-10T13:01:02.947 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: libcephfs-devel 2026-03-10T13:01:02.947 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:02.948 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:02.951 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:02.951 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:02.951 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:02.994 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: Package Arch Version Repository Size 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Removing: 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Removing dependent packages: 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Removing unused dependencies: 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Transaction Summary 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:================================================================================ 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Remove 20 Packages 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Freed space: 79 M 2026-03-10T13:01:02.996 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction check 2026-03-10T13:01:03.000 INFO:teuthology.orchestra.run.vm04.stdout:Transaction check succeeded. 2026-03-10T13:01:03.000 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction test 2026-03-10T13:01:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Transaction test succeeded. 2026-03-10T13:01:03.023 INFO:teuthology.orchestra.run.vm04.stdout:Running transaction 2026-03-10T13:01:03.067 INFO:teuthology.orchestra.run.vm04.stdout: Preparing : 1/1 2026-03-10T13:01:03.070 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T13:01:03.072 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T13:01:03.075 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T13:01:03.075 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.091 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.094 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T13:01:03.096 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T13:01:03.098 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.100 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T13:01:03.102 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T13:01:03.102 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.119 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: libcephfs-devel 2026-03-10T13:01:03.120 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:03.123 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:03.123 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:03.123 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:03.140 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:03.141 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:Remove 20 Packages 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 79 M 2026-03-10T13:01:03.142 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T13:01:03.146 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T13:01:03.146 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T13:01:03.153 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.153 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.154 INFO:teuthology.orchestra.run.vm04.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T13:01:03.154 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:03.168 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T13:01:03.168 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T13:01:03.170 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.173 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.177 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T13:01:03.182 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T13:01:03.185 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T13:01:03.188 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T13:01:03.190 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T13:01:03.192 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T13:01:03.195 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T13:01:03.210 INFO:teuthology.orchestra.run.vm04.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.210 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T13:01:03.214 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T13:01:03.216 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T13:01:03.219 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T13:01:03.219 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.232 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.234 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T13:01:03.237 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T13:01:03.238 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.240 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T13:01:03.242 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T13:01:03.242 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.258 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.258 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.258 INFO:teuthology.orchestra.run.vm09.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T13:01:03.258 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T13:01:03.274 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T13:01:03.275 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T13:01:03.275 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T13:01:03.275 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T13:01:03.275 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T13:01:03.275 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.276 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.280 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T13:01:03.284 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T13:01:03.287 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T13:01:03.290 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T13:01:03.292 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T13:01:03.294 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T13:01:03.296 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T13:01:03.314 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout:Removed: 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-10T13:01:03.324 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:03.342 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:03.343 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Remove 20 Packages 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 79 M 2026-03-10T13:01:03.344 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T13:01:03.348 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T13:01:03.348 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T13:01:03.372 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T13:01:03.373 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T13:01:03.378 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T13:01:03.417 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T13:01:03.419 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T13:01:03.421 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T13:01:03.422 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:03.425 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-10T13:01:03.425 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.441 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-10T13:01:03.443 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-10T13:01:03.445 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-10T13:01:03.447 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.449 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-10T13:01:03.451 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-10T13:01:03.451 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.467 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.468 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.468 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-10T13:01:03.468 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:03.486 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-10T13:01:03.489 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.493 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-10T13:01:03.496 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-10T13:01:03.499 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-10T13:01:03.502 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-10T13:01:03.504 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-10T13:01:03.506 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-10T13:01:03.508 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-10T13:01:03.523 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.555 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: librbd1 2026-03-10T13:01:03.555 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:03.558 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:03.559 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:03.559 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-10T13:01:03.592 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-10T13:01:03.593 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-10T13:01:03.593 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T13:01:03.645 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:03.647 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: librbd1 2026-03-10T13:01:03.647 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:03.650 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:03.650 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:03.651 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:03.766 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rados 2026-03-10T13:01:03.766 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:03.768 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:03.768 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:03.768 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:03.865 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rados 2026-03-10T13:01:03.865 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:03.868 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:03.868 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:03.868 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:03.872 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: librbd1 2026-03-10T13:01:03.873 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:03.876 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:03.876 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:03.876 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:03.949 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rgw 2026-03-10T13:01:03.949 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:03.951 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:03.952 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:03.952 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.048 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rgw 2026-03-10T13:01:04.048 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:04.051 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:04.051 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:04.051 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:04.074 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rados 2026-03-10T13:01:04.074 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:04.076 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:04.076 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:04.076 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:04.126 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-cephfs 2026-03-10T13:01:04.127 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:04.130 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:04.130 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:04.131 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.233 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-cephfs 2026-03-10T13:01:04.233 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:04.236 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:04.236 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:04.236 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:04.274 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rgw 2026-03-10T13:01:04.274 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:04.277 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:04.277 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:04.277 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:04.319 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: python3-rbd 2026-03-10T13:01:04.319 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:04.321 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:04.322 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:04.322 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.429 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: python3-rbd 2026-03-10T13:01:04.429 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:04.432 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:04.432 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:04.432 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:04.470 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-cephfs 2026-03-10T13:01:04.470 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:04.473 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:04.473 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:04.473 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:04.506 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-fuse 2026-03-10T13:01:04.506 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:04.508 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:04.509 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:04.509 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.628 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-fuse 2026-03-10T13:01:04.628 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:04.631 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:04.631 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:04.631 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:04.678 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: python3-rbd 2026-03-10T13:01:04.678 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:04.681 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:04.681 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:04.681 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:04.689 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-mirror 2026-03-10T13:01:04.689 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:04.691 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:04.691 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:04.691 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.828 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-mirror 2026-03-10T13:01:04.828 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:04.831 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:04.832 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:04.832 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:04.861 INFO:teuthology.orchestra.run.vm04.stdout:No match for argument: rbd-nbd 2026-03-10T13:01:04.862 INFO:teuthology.orchestra.run.vm04.stderr:No packages marked for removal. 2026-03-10T13:01:04.863 INFO:teuthology.orchestra.run.vm04.stdout:Dependencies resolved. 2026-03-10T13:01:04.864 INFO:teuthology.orchestra.run.vm04.stdout:Nothing to do. 2026-03-10T13:01:04.864 INFO:teuthology.orchestra.run.vm04.stdout:Complete! 2026-03-10T13:01:04.876 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-fuse 2026-03-10T13:01:04.876 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:04.878 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:04.879 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:04.879 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:04.885 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean all 2026-03-10T13:01:05.015 INFO:teuthology.orchestra.run.vm04.stdout:56 files removed 2026-03-10T13:01:05.023 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: rbd-nbd 2026-03-10T13:01:05.023 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T13:01:05.025 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T13:01:05.026 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T13:01:05.026 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T13:01:05.037 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.050 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean all 2026-03-10T13:01:05.063 DEBUG:teuthology.orchestra.run.vm04:> sudo yum clean expire-cache 2026-03-10T13:01:05.081 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-mirror 2026-03-10T13:01:05.081 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:05.083 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:05.084 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:05.084 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:05.197 INFO:teuthology.orchestra.run.vm09.stdout:56 files removed 2026-03-10T13:01:05.218 INFO:teuthology.orchestra.run.vm04.stdout:Cache was expired 2026-03-10T13:01:05.219 INFO:teuthology.orchestra.run.vm04.stdout:0 files removed 2026-03-10T13:01:05.223 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.240 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.249 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean expire-cache 2026-03-10T13:01:05.270 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: rbd-nbd 2026-03-10T13:01:05.270 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T13:01:05.272 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T13:01:05.273 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T13:01:05.273 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T13:01:05.297 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-03-10T13:01:05.415 INFO:teuthology.orchestra.run.vm09.stdout:Cache was expired 2026-03-10T13:01:05.415 INFO:teuthology.orchestra.run.vm09.stdout:0 files removed 2026-03-10T13:01:05.441 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.448 INFO:teuthology.orchestra.run.vm06.stdout:56 files removed 2026-03-10T13:01:05.474 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.501 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean expire-cache 2026-03-10T13:01:05.660 INFO:teuthology.orchestra.run.vm06.stdout:Cache was expired 2026-03-10T13:01:05.660 INFO:teuthology.orchestra.run.vm06.stdout:0 files removed 2026-03-10T13:01:05.684 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.684 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-10T13:01:05.684 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm06.local 2026-03-10T13:01:05.684 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm09.local 2026-03-10T13:01:05.684 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.684 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.684 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-10T13:01:05.714 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T13:01:05.716 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T13:01:05.717 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-10T13:01:05.782 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.783 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.786 DEBUG:teuthology.parallel:result is None 2026-03-10T13:01:05.786 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T13:01:05.788 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T13:01:05.788 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T13:01:05.826 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T13:01:05.827 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T13:01:05.840 INFO:teuthology.orchestra.run.vm04.stderr:bash: line 1: ntpq: command not found 2026-03-10T13:01:05.841 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T13:01:05.842 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:^+ srv01-nc.securepod.org 2 6 377 19 +1486us[+1486us] +/- 19ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:^* ntp3.uni-ulm.de 2 6 377 20 -1033us[-1048us] +/- 15ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:^+ 141.144.246.224 2 6 377 21 -1481us[-1497us] +/- 15ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm06.stdout:^+ web80.weingaertner-it.de 2 6 377 20 +1616us[+1616us] +/- 22ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:=============================================================================== 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:^* ntp3.uni-ulm.de 2 6 377 21 -633us[ -703us] +/- 15ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:^+ 141.144.246.224 2 6 177 17 -1523us[-1523us] +/- 16ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:^+ web80.weingaertner-it.de 2 6 377 21 +1626us[+1626us] +/- 22ms 2026-03-10T13:01:05.845 INFO:teuthology.orchestra.run.vm04.stdout:^+ srv01-nc.securepod.org 2 6 377 21 +1338us[+1338us] +/- 19ms 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:^+ web80.weingaertner-it.de 2 6 377 19 +1553us[+1553us] +/- 22ms 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:^+ srv01-nc.securepod.org 2 6 377 20 +1256us[+1334us] +/- 19ms 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:^+ ntp3.uni-ulm.de 2 6 377 19 -569us[ -491us] +/- 15ms 2026-03-10T13:01:05.846 INFO:teuthology.orchestra.run.vm09.stdout:^* 141.144.246.224 2 6 377 19 -1192us[-1114us] +/- 16ms 2026-03-10T13:01:05.846 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T13:01:05.849 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T13:01:05.849 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T13:01:05.851 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T13:01:05.853 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T13:01:05.855 INFO:teuthology.task.internal:Duration was 542.199507 seconds 2026-03-10T13:01:05.855 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T13:01:05.857 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T13:01:05.857 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T13:01:05.888 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T13:01:05.889 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T13:01:05.925 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T13:01:05.932 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T13:01:05.936 INFO:teuthology.orchestra.run.vm04.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T13:01:06.153 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T13:01:06.154 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-10T13:01:06.154 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T13:01:06.182 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-10T13:01:06.182 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T13:01:06.206 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-10T13:01:06.206 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T13:01:06.231 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T13:01:06.231 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.233 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.247 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.682 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T13:01:06.682 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:01:06.684 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:01:06.685 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T13:01:06.707 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T13:01:06.707 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T13:01:06.708 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose -- 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T13:01:06.708 INFO:teuthology.orchestra.run.vm04.stderr: /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.708 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T13:01:06.710 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T13:01:06.710 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T13:01:06.710 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 --verbose 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T13:01:06.710 INFO:teuthology.orchestra.run.vm09.stderr: -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.711 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T13:01:06.714 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T13:01:06.714 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T13:01:06.715 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T13:01:06.715 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T13:01:06.716 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T13:01:06.831 INFO:teuthology.orchestra.run.vm09.stderr: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T13:01:06.848 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T13:01:06.865 INFO:teuthology.orchestra.run.vm06.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T13:01:06.867 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T13:01:06.869 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T13:01:06.869 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T13:01:06.918 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T13:01:06.948 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T13:01:06.972 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T13:01:06.974 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:06.977 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:06.990 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:07.004 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-10T13:01:07.018 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-10T13:01:07.038 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-10T13:01:07.050 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:07.076 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T13:01:07.076 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:07.094 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T13:01:07.095 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T13:01:07.109 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T13:01:07.109 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T13:01:07.112 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T13:01:07.112 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm04 2026-03-10T13:01:07.112 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T13:01:07.147 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm06 2026-03-10T13:01:07.148 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T13:01:07.180 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/1037/remote/vm09 2026-03-10T13:01:07.181 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T13:01:07.209 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T13:01:07.209 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T13:01:07.211 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T13:01:07.223 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T13:01:07.265 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T13:01:07.267 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T13:01:07.267 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T13:01:07.270 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T13:01:07.270 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T13:01:07.272 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T13:01:07.283 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T13:01:07.286 INFO:teuthology.orchestra.run.vm04.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 13:01 /home/ubuntu/cephtest 2026-03-10T13:01:07.297 INFO:teuthology.orchestra.run.vm06.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 13:01 /home/ubuntu/cephtest 2026-03-10T13:01:07.320 INFO:teuthology.orchestra.run.vm09.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 13:01 /home/ubuntu/cephtest 2026-03-10T13:01:07.321 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T13:01:07.327 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_host_drain} duration: 542.199506521225 flavor: default owner: kyr success: true 2026-03-10T13:01:07.327 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T13:01:07.344 INFO:teuthology.run:pass