2026-03-09T19:35:58.347 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-09T19:35:58.352 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T19:35:58.373 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625 branch: squid description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_host_drain} email: null first_in_suite: false flavor: default job_id: '625' last_in_suite: false machine_type: vps name: kyr-2026-03-09_11:23:05-orch-squid-none-default-vps no_nested_subset: false os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: global: mon election default strategy: 1 mgr: debug mgr: 20 debug ms: 1 mgr/cephadm/use_agent: false mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN - mons down - mon down - out of quorum - CEPHADM_STRAY_HOST - CEPHADM_STRAY_DAEMON - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 - scontext=system_u:system_r:getty_t:s0 workunit: branch: tt-squid sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.a - osd.0 - osd.1 - - host.b - mon.b - mgr.b - osd.2 - osd.3 - - host.c - mon.c - osd.4 - osd.5 seed: 3443 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 569c3e99c9b32a51b4eaf08731c728f4513ed589 targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNviCvRex4ogen9wl/BxvykMxKv41QEaDkw3B0nGYUJXf6WbrMRWhYMXiLfH2ju7HT10IrHBrabe/hcXaQhQQkc= vm10.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAIH2bFaFYhU92zNNm9B547E2ul4rW6auk5Bi6ljQVnUdV1jk89NiyN+wz1ojBoTxCHYAPpbWqf2IG06PKlN3Cw= vm11.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBALnrTbihcHyUP0Q+K14HiuQsTdlq8/Wm1nLiIjFwpFJTb6vgj7vxzb/Z1MpToPMaxSsfklcl3aKWjjm9bRQ32Q= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install nvmetcli nvme-cli -y - install: null - cephadm: null - cephadm.shell: host.a: - "set -ex\nHOSTNAMES=$(ceph orch host ls --format json | jq -r '.[] | .hostname')\n\ for host in $HOSTNAMES; do\n # find the hostname for \"host.c\" which will\ \ have no mgr\n HAS_MGRS=$(ceph orch ps --hostname ${host} --format json |\ \ jq 'any(.daemon_type == \"mgr\")')\n if [ \"$HAS_MGRS\" == \"false\" ]; then\n\ \ HOST_C=\"${host}\"\n fi\ndone\n# One last thing to worry about before\ \ draining the host\n# is that the teuthology test tends to put the explicit\n\ # hostnames in the placement for the mon service.\n# We want to make sure we\ \ can drain without providing\n# --force and there is a check for the host being\ \ removed\n# being listed explicitly in the placements. Therefore,\n# we should\ \ remove it from the mon placement.\nceph orch ls mon --export > mon.yaml\n\ sed /\"$HOST_C\"/d mon.yaml > mon_adjusted.yaml\nceph orch apply -i mon_adjusted.yaml\n\ # now drain that host\nceph orch host drain $HOST_C --zap-osd-devices\n# wait\ \ for drain to complete\nHOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C)\n\ while [ \"$HOST_C_DAEMONS\" != \"No daemons reported\" ]; do\n sleep 15\n \ \ HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C)\ndone\n# we want to check\ \ the ability to remove the host from\n# the CRUSH map, so we should first verify\ \ the host is in\n# the CRUSH map.\nceph osd getcrushmap -o compiled-crushmap\n\ crushtool -d compiled-crushmap -o crushmap.txt\nCRUSH_MAP=$(cat crushmap.txt)\n\ if ! grep -q \"$HOST_C\" <<< \"$CRUSH_MAP\"; then\n printf \"Expected to see\ \ $HOST_C in CRUSH map. Saw:\\n\\n$CRUSH_MAP\"\n exit 1\nfi\n# If the drain\ \ was successful, we should be able to remove the\n# host without force with\ \ no issues. If there are still daemons\n# we will get a response telling us\ \ to drain the host and a\n# non-zero return code\nceph orch host rm $HOST_C\ \ --rm-crush-entry\n# verify we've successfully removed the host from the CRUSH\ \ map\nsleep 30\nceph osd getcrushmap -o compiled-crushmap\ncrushtool -d compiled-crushmap\ \ -o crushmap.txt\nCRUSH_MAP=$(cat crushmap.txt)\nif grep -q \"$HOST_C\" <<<\ \ \"$CRUSH_MAP\"; then\n printf \"Saw $HOST_C in CRUSH map after it should\ \ have been removed.\\n\\n$CRUSH_MAP\"\n exit 1\nfi\n" teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-09_11:23:05 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-09T19:35:58.373 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa; will attempt to use it 2026-03-09T19:35:58.374 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_569c3e99c9b32a51b4eaf08731c728f4513ed589/qa/tasks 2026-03-09T19:35:58.374 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-09T19:35:58.374 INFO:teuthology.task.internal:Checking packages... 2026-03-09T19:35:58.374 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-09T19:35:58.374 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-09T19:35:58.374 INFO:teuthology.packaging:ref: None 2026-03-09T19:35:58.374 INFO:teuthology.packaging:tag: None 2026-03-09T19:35:58.374 INFO:teuthology.packaging:branch: squid 2026-03-09T19:35:58.374 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:35:58.374 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-09T19:35:59.115 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-09T19:35:59.116 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-09T19:35:59.116 INFO:teuthology.task.internal:no buildpackages task found 2026-03-09T19:35:59.116 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-09T19:35:59.117 INFO:teuthology.task.internal:Saving configuration 2026-03-09T19:35:59.121 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-09T19:35:59.122 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-09T19:35:59.128 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 19:34:13.114722', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNviCvRex4ogen9wl/BxvykMxKv41QEaDkw3B0nGYUJXf6WbrMRWhYMXiLfH2ju7HT10IrHBrabe/hcXaQhQQkc='} 2026-03-09T19:35:59.132 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm10.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 19:34:13.114460', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:0a', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAIH2bFaFYhU92zNNm9B547E2ul4rW6auk5Bi6ljQVnUdV1jk89NiyN+wz1ojBoTxCHYAPpbWqf2IG06PKlN3Cw='} 2026-03-09T19:35:59.136 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm11.local', 'description': '/archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-09 19:34:13.114003', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:0b', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBALnrTbihcHyUP0Q+K14HiuQsTdlq8/Wm1nLiIjFwpFJTb6vgj7vxzb/Z1MpToPMaxSsfklcl3aKWjjm9bRQ32Q='} 2026-03-09T19:35:59.136 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-09T19:35:59.136 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['host.a', 'mon.a', 'mgr.a', 'osd.0', 'osd.1'] 2026-03-09T19:35:59.136 INFO:teuthology.task.internal:roles: ubuntu@vm10.local - ['host.b', 'mon.b', 'mgr.b', 'osd.2', 'osd.3'] 2026-03-09T19:35:59.136 INFO:teuthology.task.internal:roles: ubuntu@vm11.local - ['host.c', 'mon.c', 'osd.4', 'osd.5'] 2026-03-09T19:35:59.136 INFO:teuthology.run_tasks:Running task console_log... 2026-03-09T19:35:59.141 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-09T19:35:59.145 DEBUG:teuthology.task.console_log:vm10 does not support IPMI; excluding 2026-03-09T19:35:59.149 DEBUG:teuthology.task.console_log:vm11 does not support IPMI; excluding 2026-03-09T19:35:59.149 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fd6d3e6e170>, signals=[15]) 2026-03-09T19:35:59.149 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-09T19:35:59.150 INFO:teuthology.task.internal:Opening connections... 2026-03-09T19:35:59.150 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-09T19:35:59.150 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:35:59.209 DEBUG:teuthology.task.internal:connecting to ubuntu@vm10.local 2026-03-09T19:35:59.210 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:35:59.269 DEBUG:teuthology.task.internal:connecting to ubuntu@vm11.local 2026-03-09T19:35:59.269 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm11.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:35:59.326 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-09T19:35:59.328 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-09T19:35:59.343 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-09T19:35:59.343 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:NAME="CentOS Stream" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="9" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:ID="centos" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE="rhel fedora" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="9" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:PLATFORM_ID="platform:el9" 2026-03-09T19:35:59.397 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:ANSI_COLOR="0;31" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:LOGO="fedora-logo-icon" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://centos.org/" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T19:35:59.398 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T19:35:59.398 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-09T19:35:59.402 DEBUG:teuthology.orchestra.run.vm10:> uname -m 2026-03-09T19:35:59.417 INFO:teuthology.orchestra.run.vm10.stdout:x86_64 2026-03-09T19:35:59.417 DEBUG:teuthology.orchestra.run.vm10:> cat /etc/os-release 2026-03-09T19:35:59.471 INFO:teuthology.orchestra.run.vm10.stdout:NAME="CentOS Stream" 2026-03-09T19:35:59.471 INFO:teuthology.orchestra.run.vm10.stdout:VERSION="9" 2026-03-09T19:35:59.471 INFO:teuthology.orchestra.run.vm10.stdout:ID="centos" 2026-03-09T19:35:59.471 INFO:teuthology.orchestra.run.vm10.stdout:ID_LIKE="rhel fedora" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:VERSION_ID="9" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:PLATFORM_ID="platform:el9" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:ANSI_COLOR="0;31" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:LOGO="fedora-logo-icon" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:HOME_URL="https://centos.org/" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T19:35:59.472 INFO:teuthology.orchestra.run.vm10.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T19:35:59.472 INFO:teuthology.lock.ops:Updating vm10.local on lock server 2026-03-09T19:35:59.476 DEBUG:teuthology.orchestra.run.vm11:> uname -m 2026-03-09T19:35:59.489 INFO:teuthology.orchestra.run.vm11.stdout:x86_64 2026-03-09T19:35:59.489 DEBUG:teuthology.orchestra.run.vm11:> cat /etc/os-release 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:NAME="CentOS Stream" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:VERSION="9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:ID="centos" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:ID_LIKE="rhel fedora" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:VERSION_ID="9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:PLATFORM_ID="platform:el9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:ANSI_COLOR="0;31" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:LOGO="fedora-logo-icon" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:HOME_URL="https://centos.org/" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-09T19:35:59.544 INFO:teuthology.orchestra.run.vm11.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-09T19:35:59.544 INFO:teuthology.lock.ops:Updating vm11.local on lock server 2026-03-09T19:35:59.548 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-09T19:35:59.550 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-09T19:35:59.550 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-09T19:35:59.550 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-09T19:35:59.552 DEBUG:teuthology.orchestra.run.vm10:> test '!' -e /home/ubuntu/cephtest 2026-03-09T19:35:59.554 DEBUG:teuthology.orchestra.run.vm11:> test '!' -e /home/ubuntu/cephtest 2026-03-09T19:35:59.601 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-09T19:35:59.602 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-09T19:35:59.602 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-09T19:35:59.610 DEBUG:teuthology.orchestra.run.vm10:> test -z $(ls -A /var/lib/ceph) 2026-03-09T19:35:59.611 DEBUG:teuthology.orchestra.run.vm11:> test -z $(ls -A /var/lib/ceph) 2026-03-09T19:35:59.624 INFO:teuthology.orchestra.run.vm10.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T19:35:59.625 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T19:35:59.658 INFO:teuthology.orchestra.run.vm11.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-09T19:35:59.659 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-09T19:35:59.667 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-09T19:35:59.683 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:35:59.883 DEBUG:teuthology.orchestra.run.vm10:> test -e /ceph-qa-ready 2026-03-09T19:35:59.898 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:36:00.115 DEBUG:teuthology.orchestra.run.vm11:> test -e /ceph-qa-ready 2026-03-09T19:36:00.131 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:36:00.379 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-09T19:36:00.381 INFO:teuthology.task.internal:Creating test directory... 2026-03-09T19:36:00.381 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T19:36:00.383 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T19:36:00.385 DEBUG:teuthology.orchestra.run.vm11:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-09T19:36:00.401 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-09T19:36:00.402 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-09T19:36:00.410 INFO:teuthology.task.internal:Creating archive directory... 2026-03-09T19:36:00.410 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T19:36:00.442 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T19:36:00.444 DEBUG:teuthology.orchestra.run.vm11:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-09T19:36:00.461 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-09T19:36:00.507 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-09T19:36:00.507 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T19:36:00.523 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:36:00.523 DEBUG:teuthology.orchestra.run.vm10:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T19:36:00.537 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:36:00.537 DEBUG:teuthology.orchestra.run.vm11:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-09T19:36:00.553 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:36:00.554 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T19:36:00.565 DEBUG:teuthology.orchestra.run.vm10:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T19:36:00.579 DEBUG:teuthology.orchestra.run.vm11:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-09T19:36:00.590 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.601 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.605 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.615 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.619 INFO:teuthology.orchestra.run.vm11.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.627 INFO:teuthology.orchestra.run.vm11.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-09T19:36:00.629 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-09T19:36:00.648 INFO:teuthology.task.internal:Configuring sudo... 2026-03-09T19:36:00.648 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T19:36:00.650 DEBUG:teuthology.orchestra.run.vm10:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T19:36:00.659 DEBUG:teuthology.orchestra.run.vm11:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-09T19:36:00.692 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-09T19:36:00.713 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-09T19:36:00.714 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T19:36:00.717 DEBUG:teuthology.orchestra.run.vm10:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T19:36:00.725 DEBUG:teuthology.orchestra.run.vm11:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-09T19:36:00.746 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:36:00.793 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:36:00.852 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:36:00.852 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T19:36:00.912 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:36:00.934 DEBUG:teuthology.orchestra.run.vm10:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:36:00.991 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:36:00.991 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T19:36:01.048 DEBUG:teuthology.orchestra.run.vm11:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:36:01.070 DEBUG:teuthology.orchestra.run.vm11:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:36:01.125 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:36:01.125 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-09T19:36:01.185 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-09T19:36:01.187 DEBUG:teuthology.orchestra.run.vm10:> sudo service rsyslog restart 2026-03-09T19:36:01.189 DEBUG:teuthology.orchestra.run.vm11:> sudo service rsyslog restart 2026-03-09T19:36:01.213 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:36:01.215 INFO:teuthology.orchestra.run.vm10.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:36:01.252 INFO:teuthology.orchestra.run.vm11.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:36:01.695 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-09T19:36:01.697 INFO:teuthology.task.internal:Starting timer... 2026-03-09T19:36:01.697 INFO:teuthology.run_tasks:Running task pcp... 2026-03-09T19:36:01.700 INFO:teuthology.run_tasks:Running task selinux... 2026-03-09T19:36:01.702 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0', 'scontext=system_u:system_r:getty_t:s0']} 2026-03-09T19:36:01.702 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-09T19:36:01.702 INFO:teuthology.task.selinux:Excluding vm10: VMs are not yet supported 2026-03-09T19:36:01.702 INFO:teuthology.task.selinux:Excluding vm11: VMs are not yet supported 2026-03-09T19:36:01.702 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-09T19:36:01.702 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-09T19:36:01.702 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-09T19:36:01.702 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-09T19:36:01.704 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-09T19:36:01.704 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-09T19:36:01.706 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-09T19:36:02.320 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-09T19:36:02.325 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-09T19:36:02.326 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryays0zeld --limit vm02.local,vm10.local,vm11.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-09T19:48:15.565 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm10.local'), Remote(name='ubuntu@vm11.local')] 2026-03-09T19:48:15.566 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-09T19:48:15.566 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:48:15.627 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-09T19:48:15.706 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-09T19:48:15.706 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm10.local' 2026-03-09T19:48:15.707 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm10.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:48:15.770 DEBUG:teuthology.orchestra.run.vm10:> true 2026-03-09T19:48:15.842 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm10.local' 2026-03-09T19:48:15.842 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm11.local' 2026-03-09T19:48:15.843 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm11.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-09T19:48:15.909 DEBUG:teuthology.orchestra.run.vm11:> true 2026-03-09T19:48:15.989 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm11.local' 2026-03-09T19:48:15.990 INFO:teuthology.run_tasks:Running task clock... 2026-03-09T19:48:15.993 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-09T19:48:15.993 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T19:48:15.993 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:48:15.995 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T19:48:15.995 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:48:15.997 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-09T19:48:15.997 DEBUG:teuthology.orchestra.run.vm11:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:48:16.023 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T19:48:16.028 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T19:48:16.039 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T19:48:16.044 INFO:teuthology.orchestra.run.vm10.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T19:48:16.064 INFO:teuthology.orchestra.run.vm02.stderr:sudo: ntpd: command not found 2026-03-09T19:48:16.066 INFO:teuthology.orchestra.run.vm10.stderr:sudo: ntpd: command not found 2026-03-09T19:48:16.071 INFO:teuthology.orchestra.run.vm11.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-09T19:48:16.076 INFO:teuthology.orchestra.run.vm02.stdout:506 Cannot talk to daemon 2026-03-09T19:48:16.076 INFO:teuthology.orchestra.run.vm10.stdout:506 Cannot talk to daemon 2026-03-09T19:48:16.089 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T19:48:16.091 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T19:48:16.092 INFO:teuthology.orchestra.run.vm11.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-09T19:48:16.103 INFO:teuthology.orchestra.run.vm10.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T19:48:16.105 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T19:48:16.129 INFO:teuthology.orchestra.run.vm11.stderr:sudo: ntpd: command not found 2026-03-09T19:48:16.143 INFO:teuthology.orchestra.run.vm11.stdout:506 Cannot talk to daemon 2026-03-09T19:48:16.154 INFO:teuthology.orchestra.run.vm10.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:48:16.156 INFO:teuthology.orchestra.run.vm10.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:48:16.156 INFO:teuthology.orchestra.run.vm10.stdout:=============================================================================== 2026-03-09T19:48:16.159 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:48:16.164 INFO:teuthology.orchestra.run.vm11.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-09T19:48:16.164 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:48:16.165 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-09T19:48:16.180 INFO:teuthology.orchestra.run.vm11.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-09T19:48:16.233 INFO:teuthology.orchestra.run.vm11.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:48:16.235 INFO:teuthology.orchestra.run.vm11.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:48:16.235 INFO:teuthology.orchestra.run.vm11.stdout:=============================================================================== 2026-03-09T19:48:16.236 INFO:teuthology.run_tasks:Running task pexec... 2026-03-09T19:48:16.256 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-09T19:48:16.256 DEBUG:teuthology.orchestra.run.vm02:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T19:48:16.256 DEBUG:teuthology.orchestra.run.vm10:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T19:48:16.257 DEBUG:teuthology.orchestra.run.vm11:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-09T19:48:16.258 DEBUG:teuthology.task.pexec:ubuntu@vm10.local< sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.258 DEBUG:teuthology.task.pexec:ubuntu@vm10.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.258 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm10.local 2026-03-09T19:48:16.258 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.258 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.259 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.259 DEBUG:teuthology.task.pexec:ubuntu@vm02.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.259 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm02.local 2026-03-09T19:48:16.259 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.259 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.278 DEBUG:teuthology.task.pexec:ubuntu@vm11.local< sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.294 DEBUG:teuthology.task.pexec:ubuntu@vm11.local< sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.294 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm11.local 2026-03-09T19:48:16.294 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-09T19:48:16.294 INFO:teuthology.task.pexec:sudo dnf install nvmetcli nvme-cli -y 2026-03-09T19:48:16.462 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: nvme-cli 2026-03-09T19:48:16.463 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:48:16.465 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:48:16.466 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:48:16.466 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:48:16.479 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: nvme-cli 2026-03-09T19:48:16.479 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:48:16.483 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:48:16.483 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:48:16.483 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:48:16.526 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: nvme-cli 2026-03-09T19:48:16.526 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:48:16.529 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:48:16.531 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:48:16.532 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:48:16.878 INFO:teuthology.orchestra.run.vm10.stdout:Last metadata expiration check: 0:02:14 ago on Mon 09 Mar 2026 07:46:02 PM UTC. 2026-03-09T19:48:16.917 INFO:teuthology.orchestra.run.vm02.stdout:Last metadata expiration check: 0:11:39 ago on Mon 09 Mar 2026 07:36:37 PM UTC. 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: Package Architecture Version Repository Size 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout:Installing: 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout:Installing dependencies: 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T19:48:16.975 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:Install 6 Packages 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:Total download size: 2.3 M 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:Installed size: 11 M 2026-03-09T19:48:16.976 INFO:teuthology.orchestra.run.vm10.stdout:Downloading Packages: 2026-03-09T19:48:17.032 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: Package Architecture Version Repository Size 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Install 6 Packages 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 2.3 M 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Installed size: 11 M 2026-03-09T19:48:17.033 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-09T19:48:17.082 INFO:teuthology.orchestra.run.vm11.stdout:Last metadata expiration check: 0:00:26 ago on Mon 09 Mar 2026 07:47:51 PM UTC. 2026-03-09T19:48:17.243 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: Package Architecture Version Repository Size 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Installing: 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Installing dependencies: 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Install 6 Packages 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Total download size: 2.3 M 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Installed size: 11 M 2026-03-09T19:48:17.244 INFO:teuthology.orchestra.run.vm11.stdout:Downloading Packages: 2026-03-09T19:48:17.424 INFO:teuthology.orchestra.run.vm02.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 155 kB/s | 44 kB 00:00 2026-03-09T19:48:17.425 INFO:teuthology.orchestra.run.vm02.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 253 kB/s | 72 kB 00:00 2026-03-09T19:48:17.443 INFO:teuthology.orchestra.run.vm02.stdout:(3/6): nvme-cli-2.16-1.el9.x86_64.rpm 3.8 MB/s | 1.2 MB 00:00 2026-03-09T19:48:17.499 INFO:teuthology.orchestra.run.vm02.stdout:(4/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 15 MB/s | 837 kB 00:00 2026-03-09T19:48:17.500 INFO:teuthology.orchestra.run.vm02.stdout:(5/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.1 MB/s | 84 kB 00:00 2026-03-09T19:48:17.501 INFO:teuthology.orchestra.run.vm02.stdout:(6/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.0 MB/s | 150 kB 00:00 2026-03-09T19:48:17.501 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:48:17.501 INFO:teuthology.orchestra.run.vm02.stdout:Total 4.9 MB/s | 2.3 MB 00:00 2026-03-09T19:48:17.578 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:48:17.585 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:48:17.586 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:48:17.638 INFO:teuthology.orchestra.run.vm10.stdout:(1/6): python3-configshell-1.1.30-1.el9.noarch. 588 kB/s | 72 kB 00:00 2026-03-09T19:48:17.642 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:48:17.642 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:48:17.660 INFO:teuthology.orchestra.run.vm10.stdout:(2/6): nvmetcli-0.8-3.el9.noarch.rpm 303 kB/s | 44 kB 00:00 2026-03-09T19:48:17.724 INFO:teuthology.orchestra.run.vm10.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 979 kB/s | 84 kB 00:00 2026-03-09T19:48:17.748 INFO:teuthology.orchestra.run.vm10.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.7 MB/s | 150 kB 00:00 2026-03-09T19:48:17.849 INFO:teuthology.orchestra.run.vm10.stdout:(5/6): nvme-cli-2.16-1.el9.x86_64.rpm 3.5 MB/s | 1.2 MB 00:00 2026-03-09T19:48:17.871 INFO:teuthology.orchestra.run.vm11.stdout:(1/6): nvmetcli-0.8-3.el9.noarch.rpm 87 kB/s | 44 kB 00:00 2026-03-09T19:48:17.872 INFO:teuthology.orchestra.run.vm11.stdout:(2/6): python3-configshell-1.1.30-1.el9.noarch. 142 kB/s | 72 kB 00:00 2026-03-09T19:48:17.873 INFO:teuthology.orchestra.run.vm10.stdout:(6/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 5.5 MB/s | 837 kB 00:00 2026-03-09T19:48:17.873 INFO:teuthology.orchestra.run.vm10.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:48:17.873 INFO:teuthology.orchestra.run.vm10.stdout:Total 2.6 MB/s | 2.3 MB 00:00 2026-03-09T19:48:17.931 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:48:17.935 INFO:teuthology.orchestra.run.vm11.stdout:(3/6): python3-kmod-0.9-32.el9.x86_64.rpm 1.3 MB/s | 84 kB 00:00 2026-03-09T19:48:17.938 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:48:17.939 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:48:17.994 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:48:17.994 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:48:18.004 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:48:18.012 INFO:teuthology.orchestra.run.vm11.stdout:(4/6): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.1 MB/s | 150 kB 00:00 2026-03-09T19:48:18.026 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T19:48:18.028 INFO:teuthology.orchestra.run.vm11.stdout:(5/6): python3-urwid-2.1.2-4.el9.x86_64.rpm 8.8 MB/s | 837 kB 00:00 2026-03-09T19:48:18.040 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T19:48:18.044 INFO:teuthology.orchestra.run.vm11.stdout:(6/6): nvme-cli-2.16-1.el9.x86_64.rpm 1.7 MB/s | 1.2 MB 00:00 2026-03-09T19:48:18.044 INFO:teuthology.orchestra.run.vm11.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:48:18.044 INFO:teuthology.orchestra.run.vm11.stdout:Total 2.9 MB/s | 2.3 MB 00:00 2026-03-09T19:48:18.050 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:18.058 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:18.062 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.135 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:48:18.144 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:48:18.144 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:48:18.176 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:48:18.191 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T19:48:18.205 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T19:48:18.209 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:48:18.210 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:48:18.218 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:18.228 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:18.230 INFO:teuthology.orchestra.run.vm10.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.243 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.248 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:18.398 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:48:18.405 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.411 INFO:teuthology.orchestra.run.vm10.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:18.412 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/6 2026-03-09T19:48:18.425 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/6 2026-03-09T19:48:18.436 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:18.448 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:18.451 INFO:teuthology.orchestra.run.vm11.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.635 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:18.635 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T19:48:18.635 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:18.659 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/6 2026-03-09T19:48:18.671 INFO:teuthology.orchestra.run.vm11.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:18.756 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:18.756 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T19:48:18.756 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:19.121 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T19:48:19.121 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T19:48:19.121 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:19.121 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:19.121 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T19:48:19.122 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 6/6 2026-03-09T19:48:19.122 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-09T19:48:19.122 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:19.224 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:48:19.234 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T19:48:19.234 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T19:48:19.235 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:19.235 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:19.235 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T19:48:19.285 DEBUG:teuthology.parallel:result is None 2026-03-09T19:48:19.306 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout:Installed: 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:19.307 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:48:19.365 DEBUG:teuthology.parallel:result is None 2026-03-09T19:48:19.854 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/6 2026-03-09T19:48:19.854 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/6 2026-03-09T19:48:19.855 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/6 2026-03-09T19:48:19.855 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/6 2026-03-09T19:48:19.855 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/6 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/6 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout:Installed: 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-09T19:48:19.942 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-09T19:48:19.943 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:48:19.943 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:48:20.013 DEBUG:teuthology.parallel:result is None 2026-03-09T19:48:20.013 INFO:teuthology.run_tasks:Running task install... 2026-03-09T19:48:20.015 DEBUG:teuthology.task.install:project ceph 2026-03-09T19:48:20.016 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'}, 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T19:48:20.016 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}} 2026-03-09T19:48:20.016 INFO:teuthology.task.install:Using flavor: default 2026-03-09T19:48:20.019 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-09T19:48:20.019 INFO:teuthology.task.install:extra packages: [] 2026-03-09T19:48:20.019 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T19:48:20.019 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:48:20.020 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T19:48:20.020 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:48:20.021 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-xmltodict', 'python3-jmespath'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-xmltodict', 'python3-jmespath']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'tag': None, 'wait_for_package': False} 2026-03-09T19:48:20.021 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:48:20.611 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T19:48:20.611 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T19:48:20.675 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T19:48:20.675 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T19:48:20.702 INFO:teuthology.task.install.rpm:Pulling from https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/ 2026-03-09T19:48:20.702 INFO:teuthology.task.install.rpm:Package version is 19.2.3-678.ge911bdeb 2026-03-09T19:48:21.163 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T19:48:21.163 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:48:21.163 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T19:48:21.172 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T19:48:21.173 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:48:21.173 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T19:48:21.179 INFO:teuthology.packaging:Writing yum repo: [ceph] name=ceph packages for $basearch baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/$basearch enabled=1 gpgcheck=0 type=rpm-md [ceph-noarch] name=ceph noarch packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/noarch enabled=1 gpgcheck=0 type=rpm-md [ceph-source] name=ceph source packages baseurl=https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/SRPMS enabled=1 gpgcheck=0 type=rpm-md 2026-03-09T19:48:21.179 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:48:21.179 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-03-09T19:48:21.197 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T19:48:21.197 DEBUG:teuthology.orchestra.run.vm02:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T19:48:21.208 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T19:48:21.208 DEBUG:teuthology.orchestra.run.vm10:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T19:48:21.224 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-xmltodict, python3-jmespath on remote rpm x86_64 2026-03-09T19:48:21.224 DEBUG:teuthology.orchestra.run.vm11:> if test -f /etc/yum.repos.d/ceph.repo ; then sudo sed -i -e ':a;N;$!ba;s/enabled=1\ngpg/enabled=1\npriority=1\ngpg/g' -e 's;ref/[a-zA-Z0-9_-]*/;sha1/e911bdebe5c8faa3800735d1568fcdca65db60df/;g' /etc/yum.repos.d/ceph.repo ; fi 2026-03-09T19:48:21.272 DEBUG:teuthology.orchestra.run.vm02:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T19:48:21.274 DEBUG:teuthology.orchestra.run.vm10:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T19:48:21.300 DEBUG:teuthology.orchestra.run.vm11:> sudo touch -a /etc/yum/pluginconf.d/priorities.conf ; test -e /etc/yum/pluginconf.d/priorities.conf.orig || sudo cp -af /etc/yum/pluginconf.d/priorities.conf /etc/yum/pluginconf.d/priorities.conf.orig 2026-03-09T19:48:21.349 DEBUG:teuthology.orchestra.run.vm02:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:48:21.350 DEBUG:teuthology.orchestra.run.vm10:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:48:21.394 DEBUG:teuthology.orchestra.run.vm11:> grep check_obsoletes /etc/yum/pluginconf.d/priorities.conf && sudo sed -i 's/check_obsoletes.*0/check_obsoletes = 1/g' /etc/yum/pluginconf.d/priorities.conf || echo 'check_obsoletes = 1' | sudo tee -a /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:48:21.415 INFO:teuthology.orchestra.run.vm02.stdout:check_obsoletes = 1 2026-03-09T19:48:21.416 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-09T19:48:21.417 INFO:teuthology.orchestra.run.vm10.stdout:check_obsoletes = 1 2026-03-09T19:48:21.418 DEBUG:teuthology.orchestra.run.vm10:> sudo yum clean all 2026-03-09T19:48:21.431 INFO:teuthology.orchestra.run.vm11.stdout:check_obsoletes = 1 2026-03-09T19:48:21.434 DEBUG:teuthology.orchestra.run.vm11:> sudo yum clean all 2026-03-09T19:48:21.592 INFO:teuthology.orchestra.run.vm10.stdout:41 files removed 2026-03-09T19:48:21.602 INFO:teuthology.orchestra.run.vm02.stdout:41 files removed 2026-03-09T19:48:21.614 DEBUG:teuthology.orchestra.run.vm10:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T19:48:21.624 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T19:48:21.628 INFO:teuthology.orchestra.run.vm11.stdout:41 files removed 2026-03-09T19:48:21.649 DEBUG:teuthology.orchestra.run.vm11:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-xmltodict python3-jmespath 2026-03-09T19:48:22.931 INFO:teuthology.orchestra.run.vm10.stdout:ceph packages for x86_64 72 kB/s | 84 kB 00:01 2026-03-09T19:48:22.962 INFO:teuthology.orchestra.run.vm02.stdout:ceph packages for x86_64 71 kB/s | 84 kB 00:01 2026-03-09T19:48:23.022 INFO:teuthology.orchestra.run.vm11.stdout:ceph packages for x86_64 73 kB/s | 84 kB 00:01 2026-03-09T19:48:23.906 INFO:teuthology.orchestra.run.vm02.stdout:ceph noarch packages 13 kB/s | 12 kB 00:00 2026-03-09T19:48:23.937 INFO:teuthology.orchestra.run.vm10.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T19:48:23.991 INFO:teuthology.orchestra.run.vm11.stdout:ceph noarch packages 12 kB/s | 12 kB 00:00 2026-03-09T19:48:24.888 INFO:teuthology.orchestra.run.vm02.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T19:48:24.892 INFO:teuthology.orchestra.run.vm10.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T19:48:24.960 INFO:teuthology.orchestra.run.vm11.stdout:ceph source packages 2.0 kB/s | 1.9 kB 00:00 2026-03-09T19:48:26.111 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - BaseOS 7.4 MB/s | 8.9 MB 00:01 2026-03-09T19:48:27.052 INFO:teuthology.orchestra.run.vm11.stdout:CentOS Stream 9 - BaseOS 4.3 MB/s | 8.9 MB 00:02 2026-03-09T19:48:28.053 INFO:teuthology.orchestra.run.vm10.stdout:CentOS Stream 9 - BaseOS 2.8 MB/s | 8.9 MB 00:03 2026-03-09T19:48:29.942 INFO:teuthology.orchestra.run.vm10.stdout:CentOS Stream 9 - AppStream 22 MB/s | 27 MB 00:01 2026-03-09T19:48:32.744 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - AppStream 4.6 MB/s | 27 MB 00:05 2026-03-09T19:48:34.727 INFO:teuthology.orchestra.run.vm10.stdout:CentOS Stream 9 - CRB 4.1 MB/s | 8.0 MB 00:01 2026-03-09T19:48:36.223 INFO:teuthology.orchestra.run.vm10.stdout:CentOS Stream 9 - Extras packages 33 kB/s | 20 kB 00:00 2026-03-09T19:48:36.856 INFO:teuthology.orchestra.run.vm10.stdout:Extra Packages for Enterprise Linux 38 MB/s | 20 MB 00:00 2026-03-09T19:48:37.509 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - CRB 4.1 MB/s | 8.0 MB 00:01 2026-03-09T19:48:39.783 INFO:teuthology.orchestra.run.vm02.stdout:CentOS Stream 9 - Extras packages 14 kB/s | 20 kB 00:01 2026-03-09T19:48:40.825 INFO:teuthology.orchestra.run.vm02.stdout:Extra Packages for Enterprise Linux 21 MB/s | 20 MB 00:00 2026-03-09T19:48:41.644 INFO:teuthology.orchestra.run.vm10.stdout:lab-extras 62 kB/s | 50 kB 00:00 2026-03-09T19:48:43.160 INFO:teuthology.orchestra.run.vm10.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:48:43.161 INFO:teuthology.orchestra.run.vm10.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:48:43.165 INFO:teuthology.orchestra.run.vm10.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T19:48:43.166 INFO:teuthology.orchestra.run.vm10.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T19:48:43.194 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout:====================================================================================== 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout:====================================================================================== 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout:Installing: 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T19:48:43.198 INFO:teuthology.orchestra.run.vm10.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout:Upgrading: 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout:Installing dependencies: 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T19:48:43.199 INFO:teuthology.orchestra.run.vm10.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T19:48:43.200 INFO:teuthology.orchestra.run.vm10.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Installing weak dependencies: 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:====================================================================================== 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Install 134 Packages 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Upgrade 2 Packages 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Total download size: 210 M 2026-03-09T19:48:43.201 INFO:teuthology.orchestra.run.vm10.stdout:Downloading Packages: 2026-03-09T19:48:45.377 INFO:teuthology.orchestra.run.vm10.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 14 kB/s | 6.5 kB 00:00 2026-03-09T19:48:45.550 INFO:teuthology.orchestra.run.vm02.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-03-09T19:48:46.421 INFO:teuthology.orchestra.run.vm10.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.1 MB/s | 1.2 MB 00:01 2026-03-09T19:48:46.594 INFO:teuthology.orchestra.run.vm10.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 841 kB/s | 145 kB 00:00 2026-03-09T19:48:47.091 INFO:teuthology.orchestra.run.vm02.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:48:47.091 INFO:teuthology.orchestra.run.vm02.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:48:47.096 INFO:teuthology.orchestra.run.vm02.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T19:48:47.096 INFO:teuthology.orchestra.run.vm02.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T19:48:47.126 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout:====================================================================================== 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout:====================================================================================== 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout:Upgrading: 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T19:48:47.131 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T19:48:47.132 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:Installing weak dependencies: 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:====================================================================================== 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:Install 134 Packages 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:Upgrade 2 Packages 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:48:47.133 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 210 M 2026-03-09T19:48:47.134 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-03-09T19:48:47.603 INFO:teuthology.orchestra.run.vm10.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 2.4 MB/s | 2.4 MB 00:01 2026-03-09T19:48:47.833 INFO:teuthology.orchestra.run.vm10.stdout:(5/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.9 MB/s | 5.5 MB 00:02 2026-03-09T19:48:48.025 INFO:teuthology.orchestra.run.vm10.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 2.6 MB/s | 1.1 MB 00:00 2026-03-09T19:48:48.664 INFO:teuthology.orchestra.run.vm02.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 11 kB/s | 6.5 kB 00:00 2026-03-09T19:48:49.272 INFO:teuthology.orchestra.run.vm10.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.3 MB/s | 4.7 MB 00:01 2026-03-09T19:48:49.889 INFO:teuthology.orchestra.run.vm02.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 962 kB/s | 1.2 MB 00:01 2026-03-09T19:48:50.045 INFO:teuthology.orchestra.run.vm02.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 933 kB/s | 145 kB 00:00 2026-03-09T19:48:51.174 INFO:teuthology.orchestra.run.vm10.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 5.6 MB/s | 11 MB 00:01 2026-03-09T19:48:51.250 INFO:teuthology.orchestra.run.vm02.stdout:(4/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 2.0 MB/s | 2.4 MB 00:01 2026-03-09T19:48:51.333 INFO:teuthology.orchestra.run.vm10.stdout:(9/136): ceph-selinux-19.2.3-678.ge911bdeb.el9. 158 kB/s | 25 kB 00:00 2026-03-09T19:48:51.476 INFO:teuthology.orchestra.run.vm10.stdout:(10/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86 4.9 MB/s | 17 MB 00:03 2026-03-09T19:48:51.634 INFO:teuthology.orchestra.run.vm10.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 214 kB/s | 34 kB 00:00 2026-03-09T19:48:51.710 INFO:teuthology.orchestra.run.vm02.stdout:(5/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 2.3 MB/s | 1.1 MB 00:00 2026-03-09T19:48:51.742 INFO:teuthology.orchestra.run.vm02.stdout:(6/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 1.5 MB/s | 5.5 MB 00:03 2026-03-09T19:48:51.832 INFO:teuthology.orchestra.run.vm10.stdout:(12/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 4.9 MB/s | 1.0 MB 00:00 2026-03-09T19:48:51.990 INFO:teuthology.orchestra.run.vm10.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.0 MB/s | 163 kB 00:00 2026-03-09T19:48:52.160 INFO:teuthology.orchestra.run.vm10.stdout:(14/136): librados-devel-19.2.3-678.ge911bdeb.e 746 kB/s | 127 kB 00:00 2026-03-09T19:48:52.328 INFO:teuthology.orchestra.run.vm10.stdout:(15/136): libradosstriper1-19.2.3-678.ge911bdeb 2.9 MB/s | 503 kB 00:00 2026-03-09T19:48:52.991 INFO:teuthology.orchestra.run.vm10.stdout:(16/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 8.1 MB/s | 5.4 MB 00:00 2026-03-09T19:48:53.097 INFO:teuthology.orchestra.run.vm02.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 3.4 MB/s | 4.7 MB 00:01 2026-03-09T19:48:53.153 INFO:teuthology.orchestra.run.vm10.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 280 kB/s | 45 kB 00:00 2026-03-09T19:48:53.312 INFO:teuthology.orchestra.run.vm10.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 894 kB/s | 142 kB 00:00 2026-03-09T19:48:53.479 INFO:teuthology.orchestra.run.vm10.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 989 kB/s | 165 kB 00:00 2026-03-09T19:48:53.626 INFO:teuthology.orchestra.run.vm10.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.1 MB/s | 323 kB 00:00 2026-03-09T19:48:53.807 INFO:teuthology.orchestra.run.vm10.stdout:(21/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 1.6 MB/s | 303 kB 00:00 2026-03-09T19:48:53.951 INFO:teuthology.orchestra.run.vm10.stdout:(22/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 691 kB/s | 100 kB 00:00 2026-03-09T19:48:54.114 INFO:teuthology.orchestra.run.vm10.stdout:(23/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 524 kB/s | 85 kB 00:00 2026-03-09T19:48:54.435 INFO:teuthology.orchestra.run.vm10.stdout:(24/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 9.7 MB/s | 3.1 MB 00:00 2026-03-09T19:48:54.582 INFO:teuthology.orchestra.run.vm10.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.1 MB/s | 171 kB 00:00 2026-03-09T19:48:54.726 INFO:teuthology.orchestra.run.vm10.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 216 kB/s | 31 kB 00:00 2026-03-09T19:48:54.869 INFO:teuthology.orchestra.run.vm02.stdout:(8/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9. 6.1 MB/s | 11 MB 00:01 2026-03-09T19:48:54.881 INFO:teuthology.orchestra.run.vm10.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 971 kB/s | 150 kB 00:00 2026-03-09T19:48:55.180 INFO:teuthology.orchestra.run.vm02.stdout:(9/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86_ 5.0 MB/s | 17 MB 00:03 2026-03-09T19:48:55.231 INFO:teuthology.orchestra.run.vm02.stdout:(10/136): ceph-selinux-19.2.3-678.ge911bdeb.el9 69 kB/s | 25 kB 00:00 2026-03-09T19:48:55.231 INFO:teuthology.orchestra.run.vm10.stdout:(28/136): ceph-test-19.2.3-678.ge911bdeb.el9.x8 13 MB/s | 50 MB 00:03 2026-03-09T19:48:55.452 INFO:teuthology.orchestra.run.vm02.stdout:(11/136): ceph-common-19.2.3-678.ge911bdeb.el9. 2.9 MB/s | 22 MB 00:07 2026-03-09T19:48:55.511 INFO:teuthology.orchestra.run.vm10.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 6.0 MB/s | 3.8 MB 00:00 2026-03-09T19:48:55.518 INFO:teuthology.orchestra.run.vm02.stdout:(12/136): libcephfs-devel-19.2.3-678.ge911bdeb. 117 kB/s | 34 kB 00:00 2026-03-09T19:48:55.652 INFO:teuthology.orchestra.run.vm02.stdout:(13/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.2 MB/s | 163 kB 00:00 2026-03-09T19:48:55.657 INFO:teuthology.orchestra.run.vm10.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.7 MB/s | 253 kB 00:00 2026-03-09T19:48:55.753 INFO:teuthology.orchestra.run.vm02.stdout:(14/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 3.2 MB/s | 1.0 MB 00:00 2026-03-09T19:48:55.783 INFO:teuthology.orchestra.run.vm02.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 965 kB/s | 127 kB 00:00 2026-03-09T19:48:55.806 INFO:teuthology.orchestra.run.vm10.stdout:(31/136): ceph-mgr-diskprediction-local-19.2.3- 13 MB/s | 7.4 MB 00:00 2026-03-09T19:48:55.807 INFO:teuthology.orchestra.run.vm10.stdout:(32/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 328 kB/s | 49 kB 00:00 2026-03-09T19:48:55.901 INFO:teuthology.orchestra.run.vm02.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 3.3 MB/s | 503 kB 00:00 2026-03-09T19:48:55.944 INFO:teuthology.orchestra.run.vm10.stdout:(33/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.1 MB/s | 299 kB 00:00 2026-03-09T19:48:55.945 INFO:teuthology.orchestra.run.vm10.stdout:(34/136): ceph-prometheus-alerts-19.2.3-678.ge9 120 kB/s | 17 kB 00:00 2026-03-09T19:48:56.043 INFO:teuthology.orchestra.run.vm02.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 318 kB/s | 45 kB 00:00 2026-03-09T19:48:56.087 INFO:teuthology.orchestra.run.vm10.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 5.3 MB/s | 769 kB 00:00 2026-03-09T19:48:56.187 INFO:teuthology.orchestra.run.vm02.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 992 kB/s | 142 kB 00:00 2026-03-09T19:48:56.231 INFO:teuthology.orchestra.run.vm10.stdout:(36/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 1.2 MB/s | 351 kB 00:00 2026-03-09T19:48:56.231 INFO:teuthology.orchestra.run.vm10.stdout:(37/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 281 kB/s | 40 kB 00:00 2026-03-09T19:48:56.292 INFO:teuthology.orchestra.run.vm10.stdout:(38/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.1 MB/s | 72 kB 00:00 2026-03-09T19:48:56.330 INFO:teuthology.orchestra.run.vm02.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.1 MB/s | 165 kB 00:00 2026-03-09T19:48:56.360 INFO:teuthology.orchestra.run.vm10.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 2.7 MB/s | 184 kB 00:00 2026-03-09T19:48:56.406 INFO:teuthology.orchestra.run.vm10.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 727 kB/s | 33 kB 00:00 2026-03-09T19:48:56.447 INFO:teuthology.orchestra.run.vm10.stdout:(41/136): libgfortran-11.5.0-14.el9.x86_64.rpm 3.6 MB/s | 794 kB 00:00 2026-03-09T19:48:56.462 INFO:teuthology.orchestra.run.vm10.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 1.6 MB/s | 93 kB 00:00 2026-03-09T19:48:56.476 INFO:teuthology.orchestra.run.vm02.stdout:(20/136): python3-rados-19.2.3-678.ge911bdeb.el 2.2 MB/s | 323 kB 00:00 2026-03-09T19:48:56.524 INFO:teuthology.orchestra.run.vm10.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 3.2 MB/s | 253 kB 00:00 2026-03-09T19:48:56.594 INFO:teuthology.orchestra.run.vm10.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 1.5 MB/s | 106 kB 00:00 2026-03-09T19:48:56.594 INFO:teuthology.orchestra.run.vm02.stdout:(21/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 6.7 MB/s | 5.4 MB 00:00 2026-03-09T19:48:56.620 INFO:teuthology.orchestra.run.vm10.stdout:(45/136): python3-cryptography-36.0.1-5.el9.x86 7.9 MB/s | 1.2 MB 00:00 2026-03-09T19:48:56.622 INFO:teuthology.orchestra.run.vm02.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 2.0 MB/s | 303 kB 00:00 2026-03-09T19:48:56.657 INFO:teuthology.orchestra.run.vm10.stdout:(46/136): python3-pycparser-2.20-6.el9.noarch.r 2.1 MB/s | 135 kB 00:00 2026-03-09T19:48:56.691 INFO:teuthology.orchestra.run.vm10.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 1.7 MB/s | 126 kB 00:00 2026-03-09T19:48:56.726 INFO:teuthology.orchestra.run.vm02.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 759 kB/s | 100 kB 00:00 2026-03-09T19:48:56.730 INFO:teuthology.orchestra.run.vm10.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 2.9 MB/s | 218 kB 00:00 2026-03-09T19:48:56.763 INFO:teuthology.orchestra.run.vm10.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 2.5 MB/s | 182 kB 00:00 2026-03-09T19:48:56.766 INFO:teuthology.orchestra.run.vm02.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 591 kB/s | 85 kB 00:00 2026-03-09T19:48:56.832 INFO:teuthology.orchestra.run.vm10.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 2.5 MB/s | 266 kB 00:00 2026-03-09T19:48:56.835 INFO:teuthology.orchestra.run.vm10.stdout:(51/136): boost-program-options-1.75.0-13.el9.x 1.4 MB/s | 104 kB 00:00 2026-03-09T19:48:56.882 INFO:teuthology.orchestra.run.vm10.stdout:(52/136): flexiblas-3.0.4-9.el9.x86_64.rpm 602 kB/s | 30 kB 00:00 2026-03-09T19:48:56.901 INFO:teuthology.orchestra.run.vm10.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 782 kB/s | 15 kB 00:00 2026-03-09T19:48:56.911 INFO:teuthology.orchestra.run.vm02.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.2 MB/s | 171 kB 00:00 2026-03-09T19:48:56.932 INFO:teuthology.orchestra.run.vm10.stdout:(54/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 31 MB/s | 3.0 MB 00:00 2026-03-09T19:48:56.948 INFO:teuthology.orchestra.run.vm10.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 9.8 MB/s | 160 kB 00:00 2026-03-09T19:48:56.954 INFO:teuthology.orchestra.run.vm10.stdout:(56/136): libnbd-1.20.3-4.el9.x86_64.rpm 3.0 MB/s | 164 kB 00:00 2026-03-09T19:48:56.963 INFO:teuthology.orchestra.run.vm10.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 3.0 MB/s | 45 kB 00:00 2026-03-09T19:48:56.980 INFO:teuthology.orchestra.run.vm10.stdout:(58/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 14 MB/s | 246 kB 00:00 2026-03-09T19:48:57.000 INFO:teuthology.orchestra.run.vm10.stdout:(59/136): libxslt-1.1.34-12.el9.x86_64.rpm 12 MB/s | 233 kB 00:00 2026-03-09T19:48:57.005 INFO:teuthology.orchestra.run.vm10.stdout:(60/136): librdkafka-1.6.1-102.el9.x86_64.rpm 13 MB/s | 662 kB 00:00 2026-03-09T19:48:57.018 INFO:teuthology.orchestra.run.vm10.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 16 MB/s | 292 kB 00:00 2026-03-09T19:48:57.022 INFO:teuthology.orchestra.run.vm10.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 11 MB/s | 188 kB 00:00 2026-03-09T19:48:57.034 INFO:teuthology.orchestra.run.vm10.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 2.7 MB/s | 42 kB 00:00 2026-03-09T19:48:57.106 INFO:teuthology.orchestra.run.vm10.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 14 MB/s | 1.0 MB 00:00 2026-03-09T19:48:57.165 INFO:teuthology.orchestra.run.vm10.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 37 MB/s | 5.3 MB 00:00 2026-03-09T19:48:57.183 INFO:teuthology.orchestra.run.vm10.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 14 MB/s | 244 kB 00:00 2026-03-09T19:48:57.200 INFO:teuthology.orchestra.run.vm10.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 14 MB/s | 249 kB 00:00 2026-03-09T19:48:57.215 INFO:teuthology.orchestra.run.vm10.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 3.1 MB/s | 48 kB 00:00 2026-03-09T19:48:57.232 INFO:teuthology.orchestra.run.vm10.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 11 MB/s | 177 kB 00:00 2026-03-09T19:48:57.249 INFO:teuthology.orchestra.run.vm10.stdout:(70/136): python3-mako-1.1.4-6.el9.noarch.rpm 9.6 MB/s | 172 kB 00:00 2026-03-09T19:48:57.264 INFO:teuthology.orchestra.run.vm10.stdout:(71/136): python3-markupsafe-1.1.1-12.el9.x86_6 2.3 MB/s | 35 kB 00:00 2026-03-09T19:48:57.291 INFO:teuthology.orchestra.run.vm10.stdout:(72/136): python3-babel-2.9.1-2.el9.noarch.rpm 32 MB/s | 6.0 MB 00:00 2026-03-09T19:48:57.356 INFO:teuthology.orchestra.run.vm10.stdout:(73/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 6.7 MB/s | 442 kB 00:00 2026-03-09T19:48:57.377 INFO:teuthology.orchestra.run.vm10.stdout:(74/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 55 MB/s | 6.1 MB 00:00 2026-03-09T19:48:57.377 INFO:teuthology.orchestra.run.vm10.stdout:(75/136): python3-packaging-20.9-5.el9.noarch.r 3.5 MB/s | 77 kB 00:00 2026-03-09T19:48:57.394 INFO:teuthology.orchestra.run.vm10.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 16 MB/s | 267 kB 00:00 2026-03-09T19:48:57.395 INFO:teuthology.orchestra.run.vm10.stdout:(77/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 8.8 MB/s | 157 kB 00:00 2026-03-09T19:48:57.411 INFO:teuthology.orchestra.run.vm10.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 16 MB/s | 277 kB 00:00 2026-03-09T19:48:57.411 INFO:teuthology.orchestra.run.vm10.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 3.2 MB/s | 54 kB 00:00 2026-03-09T19:48:57.428 INFO:teuthology.orchestra.run.vm10.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 2.5 MB/s | 42 kB 00:00 2026-03-09T19:48:57.447 INFO:teuthology.orchestra.run.vm10.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 13 MB/s | 240 kB 00:00 2026-03-09T19:48:57.462 INFO:teuthology.orchestra.run.vm10.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 2.5 MB/s | 37 kB 00:00 2026-03-09T19:48:57.506 INFO:teuthology.orchestra.run.vm10.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.5 MB/s | 66 kB 00:00 2026-03-09T19:48:57.582 INFO:teuthology.orchestra.run.vm10.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 3.9 MB/s | 303 kB 00:00 2026-03-09T19:48:57.611 INFO:teuthology.orchestra.run.vm10.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.2 MB/s | 64 kB 00:00 2026-03-09T19:48:57.716 INFO:teuthology.orchestra.run.vm10.stdout:(86/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 63 MB/s | 19 MB 00:00 2026-03-09T19:48:57.939 INFO:teuthology.orchestra.run.vm10.stdout:(87/136): lua-devel-5.4.4-4.el9.x86_64.rpm 68 kB/s | 22 kB 00:00 2026-03-09T19:48:57.953 INFO:teuthology.orchestra.run.vm10.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 39 MB/s | 551 kB 00:00 2026-03-09T19:48:57.960 INFO:teuthology.orchestra.run.vm10.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 47 MB/s | 308 kB 00:00 2026-03-09T19:48:57.962 INFO:teuthology.orchestra.run.vm10.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 8.8 MB/s | 19 kB 00:00 2026-03-09T19:48:58.040 INFO:teuthology.orchestra.run.vm10.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 57 MB/s | 4.4 MB 00:00 2026-03-09T19:48:58.053 INFO:teuthology.orchestra.run.vm10.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 1.9 MB/s | 25 kB 00:00 2026-03-09T19:48:58.054 INFO:teuthology.orchestra.run.vm02.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 27 kB/s | 31 kB 00:01 2026-03-09T19:48:58.056 INFO:teuthology.orchestra.run.vm10.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 18 MB/s | 49 kB 00:00 2026-03-09T19:48:58.059 INFO:teuthology.orchestra.run.vm10.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 24 MB/s | 67 kB 00:00 2026-03-09T19:48:58.063 INFO:teuthology.orchestra.run.vm10.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 40 MB/s | 151 kB 00:00 2026-03-09T19:48:58.077 INFO:teuthology.orchestra.run.vm10.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 56 MB/s | 838 kB 00:00 2026-03-09T19:48:58.086 INFO:teuthology.orchestra.run.vm10.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 62 MB/s | 548 kB 00:00 2026-03-09T19:48:58.089 INFO:teuthology.orchestra.run.vm10.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 13 MB/s | 29 kB 00:00 2026-03-09T19:48:58.091 INFO:teuthology.orchestra.run.vm10.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 22 MB/s | 60 kB 00:00 2026-03-09T19:48:58.094 INFO:teuthology.orchestra.run.vm10.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 16 MB/s | 43 kB 00:00 2026-03-09T19:48:58.097 INFO:teuthology.orchestra.run.vm10.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 13 MB/s | 32 kB 00:00 2026-03-09T19:48:58.099 INFO:teuthology.orchestra.run.vm10.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 6.2 MB/s | 14 kB 00:00 2026-03-09T19:48:58.103 INFO:teuthology.orchestra.run.vm10.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 43 MB/s | 173 kB 00:00 2026-03-09T19:48:58.109 INFO:teuthology.orchestra.run.vm10.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 57 MB/s | 358 kB 00:00 2026-03-09T19:48:58.114 INFO:teuthology.orchestra.run.vm10.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 53 MB/s | 254 kB 00:00 2026-03-09T19:48:58.142 INFO:teuthology.orchestra.run.vm10.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 73 MB/s | 2.0 MB 00:00 2026-03-09T19:48:58.146 INFO:teuthology.orchestra.run.vm10.stdout:(107/136): protobuf-compiler-3.14.0-17.el9.x86_ 2.0 MB/s | 862 kB 00:00 2026-03-09T19:48:58.147 INFO:teuthology.orchestra.run.vm10.stdout:(108/136): python3-grpcio-tools-1.46.7-10.el9.x 32 MB/s | 144 kB 00:00 2026-03-09T19:48:58.148 INFO:teuthology.orchestra.run.vm11.stdout:CentOS Stream 9 - AppStream 922 kB/s | 27 MB 00:30 2026-03-09T19:48:58.149 INFO:teuthology.orchestra.run.vm10.stdout:(109/136): python3-jaraco-8.2.1-3.el9.noarch.rp 3.3 MB/s | 11 kB 00:00 2026-03-09T19:48:58.150 INFO:teuthology.orchestra.run.vm10.stdout:(110/136): python3-jaraco-classes-3.2.1-5.el9.n 6.0 MB/s | 18 kB 00:00 2026-03-09T19:48:58.152 INFO:teuthology.orchestra.run.vm10.stdout:(111/136): python3-jaraco-collections-3.0.0-8.e 9.7 MB/s | 23 kB 00:00 2026-03-09T19:48:58.153 INFO:teuthology.orchestra.run.vm10.stdout:(112/136): python3-jaraco-context-6.0.1-3.el9.n 8.2 MB/s | 20 kB 00:00 2026-03-09T19:48:58.154 INFO:teuthology.orchestra.run.vm10.stdout:(113/136): python3-jaraco-functools-3.5.0-2.el9 9.0 MB/s | 19 kB 00:00 2026-03-09T19:48:58.155 INFO:teuthology.orchestra.run.vm10.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 12 MB/s | 26 kB 00:00 2026-03-09T19:48:58.157 INFO:teuthology.orchestra.run.vm10.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 20 MB/s | 46 kB 00:00 2026-03-09T19:48:58.162 INFO:teuthology.orchestra.run.vm10.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 20 MB/s | 79 kB 00:00 2026-03-09T19:48:58.169 INFO:teuthology.orchestra.run.vm10.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 8.3 MB/s | 58 kB 00:00 2026-03-09T19:48:58.173 INFO:teuthology.orchestra.run.vm10.stdout:(118/136): python3-kubernetes-26.1.0-3.el9.noar 56 MB/s | 1.0 MB 00:00 2026-03-09T19:48:58.174 INFO:teuthology.orchestra.run.vm10.stdout:(119/136): python3-pecan-1.4.2-3.el9.noarch.rpm 48 MB/s | 272 kB 00:00 2026-03-09T19:48:58.175 INFO:teuthology.orchestra.run.vm10.stdout:(120/136): python3-portend-3.1.0-2.el9.noarch.r 7.9 MB/s | 16 kB 00:00 2026-03-09T19:48:58.178 INFO:teuthology.orchestra.run.vm10.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 26 MB/s | 90 kB 00:00 2026-03-09T19:48:58.178 INFO:teuthology.orchestra.run.vm10.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 9.0 MB/s | 31 kB 00:00 2026-03-09T19:48:58.182 INFO:teuthology.orchestra.run.vm10.stdout:(123/136): python3-routes-2.5.1-5.el9.noarch.rp 43 MB/s | 188 kB 00:00 2026-03-09T19:48:58.183 INFO:teuthology.orchestra.run.vm10.stdout:(124/136): python3-rsa-4.9-2.el9.noarch.rpm 13 MB/s | 59 kB 00:00 2026-03-09T19:48:58.185 INFO:teuthology.orchestra.run.vm10.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 13 MB/s | 36 kB 00:00 2026-03-09T19:48:58.187 INFO:teuthology.orchestra.run.vm10.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 21 MB/s | 86 kB 00:00 2026-03-09T19:48:58.191 INFO:teuthology.orchestra.run.vm10.stdout:(127/136): python3-webob-1.8.8-2.el9.noarch.rpm 36 MB/s | 230 kB 00:00 2026-03-09T19:48:58.192 INFO:teuthology.orchestra.run.vm10.stdout:(128/136): python3-websocket-client-1.2.3-2.el9 18 MB/s | 90 kB 00:00 2026-03-09T19:48:58.195 INFO:teuthology.orchestra.run.vm10.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 8.9 MB/s | 22 kB 00:00 2026-03-09T19:48:58.197 INFO:teuthology.orchestra.run.vm10.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 7.9 MB/s | 20 kB 00:00 2026-03-09T19:48:58.202 INFO:teuthology.orchestra.run.vm10.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 43 MB/s | 191 kB 00:00 2026-03-09T19:48:58.220 INFO:teuthology.orchestra.run.vm10.stdout:(132/136): python3-werkzeug-2.0.3-3.el9.1.noarc 15 MB/s | 427 kB 00:00 2026-03-09T19:48:58.225 INFO:teuthology.orchestra.run.vm10.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 71 MB/s | 1.6 MB 00:00 2026-03-09T19:48:58.486 INFO:teuthology.orchestra.run.vm10.stdout:(134/136): ceph-common-19.2.3-678.ge911bdeb.el9 1.6 MB/s | 22 MB 00:13 2026-03-09T19:48:58.764 INFO:teuthology.orchestra.run.vm02.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 212 kB/s | 150 kB 00:00 2026-03-09T19:48:58.881 INFO:teuthology.orchestra.run.vm02.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 1.4 MB/s | 3.1 MB 00:02 2026-03-09T19:48:59.550 INFO:teuthology.orchestra.run.vm10.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 2.4 MB/s | 3.2 MB 00:01 2026-03-09T19:48:59.827 INFO:teuthology.orchestra.run.vm10.stdout:(136/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.1 MB/s | 3.4 MB 00:01 2026-03-09T19:48:59.829 INFO:teuthology.orchestra.run.vm10.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:48:59.830 INFO:teuthology.orchestra.run.vm10.stdout:Total 13 MB/s | 210 MB 00:16 2026-03-09T19:48:59.914 INFO:teuthology.orchestra.run.vm02.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 3.3 MB/s | 3.8 MB 00:01 2026-03-09T19:48:59.958 INFO:teuthology.orchestra.run.vm02.stdout:(30/136): ceph-mgr-diskprediction-local-19.2.3- 6.9 MB/s | 7.4 MB 00:01 2026-03-09T19:49:00.059 INFO:teuthology.orchestra.run.vm02.stdout:(31/136): ceph-mgr-modules-core-19.2.3-678.ge91 1.7 MB/s | 253 kB 00:00 2026-03-09T19:49:00.091 INFO:teuthology.orchestra.run.vm02.stdout:(32/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 370 kB/s | 49 kB 00:00 2026-03-09T19:49:00.202 INFO:teuthology.orchestra.run.vm02.stdout:(33/136): ceph-prometheus-alerts-19.2.3-678.ge9 117 kB/s | 17 kB 00:00 2026-03-09T19:49:00.227 INFO:teuthology.orchestra.run.vm02.stdout:(34/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 2.1 MB/s | 299 kB 00:00 2026-03-09T19:49:00.471 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:49:00.491 INFO:teuthology.orchestra.run.vm02.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 2.6 MB/s | 769 kB 00:00 2026-03-09T19:49:00.527 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:49:00.527 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:49:00.618 INFO:teuthology.orchestra.run.vm02.stdout:(36/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 321 kB/s | 40 kB 00:00 2026-03-09T19:49:00.676 INFO:teuthology.orchestra.run.vm02.stdout:(37/136): libconfig-1.7.2-9.el9.x86_64.rpm 1.2 MB/s | 72 kB 00:00 2026-03-09T19:49:00.702 INFO:teuthology.orchestra.run.vm02.stdout:(38/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 740 kB/s | 351 kB 00:00 2026-03-09T19:49:00.734 INFO:teuthology.orchestra.run.vm02.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 5.7 MB/s | 184 kB 00:00 2026-03-09T19:49:00.766 INFO:teuthology.orchestra.run.vm02.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 1.0 MB/s | 33 kB 00:00 2026-03-09T19:49:00.768 INFO:teuthology.orchestra.run.vm02.stdout:(41/136): libgfortran-11.5.0-14.el9.x86_64.rpm 8.4 MB/s | 794 kB 00:00 2026-03-09T19:49:00.796 INFO:teuthology.orchestra.run.vm02.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 3.0 MB/s | 93 kB 00:00 2026-03-09T19:49:00.801 INFO:teuthology.orchestra.run.vm02.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 7.6 MB/s | 253 kB 00:00 2026-03-09T19:49:00.832 INFO:teuthology.orchestra.run.vm02.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 3.4 MB/s | 106 kB 00:00 2026-03-09T19:49:00.863 INFO:teuthology.orchestra.run.vm02.stdout:(45/136): python3-pycparser-2.20-6.el9.noarch.r 4.3 MB/s | 135 kB 00:00 2026-03-09T19:49:00.885 INFO:teuthology.orchestra.run.vm02.stdout:(46/136): python3-cryptography-36.0.1-5.el9.x86 14 MB/s | 1.2 MB 00:00 2026-03-09T19:49:00.894 INFO:teuthology.orchestra.run.vm02.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 4.0 MB/s | 126 kB 00:00 2026-03-09T19:49:00.918 INFO:teuthology.orchestra.run.vm02.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 6.6 MB/s | 218 kB 00:00 2026-03-09T19:49:00.926 INFO:teuthology.orchestra.run.vm02.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 5.7 MB/s | 182 kB 00:00 2026-03-09T19:49:00.950 INFO:teuthology.orchestra.run.vm02.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 8.1 MB/s | 266 kB 00:00 2026-03-09T19:49:00.957 INFO:teuthology.orchestra.run.vm02.stdout:(51/136): boost-program-options-1.75.0-13.el9.x 3.3 MB/s | 104 kB 00:00 2026-03-09T19:49:00.979 INFO:teuthology.orchestra.run.vm02.stdout:(52/136): flexiblas-3.0.4-9.el9.x86_64.rpm 1.0 MB/s | 30 kB 00:00 2026-03-09T19:49:01.009 INFO:teuthology.orchestra.run.vm02.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 507 kB/s | 15 kB 00:00 2026-03-09T19:49:01.040 INFO:teuthology.orchestra.run.vm02.stdout:(54/136): libnbd-1.20.3-4.el9.x86_64.rpm 5.2 MB/s | 164 kB 00:00 2026-03-09T19:49:01.071 INFO:teuthology.orchestra.run.vm02.stdout:(55/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 5.0 MB/s | 160 kB 00:00 2026-03-09T19:49:01.085 INFO:teuthology.orchestra.run.vm02.stdout:(56/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 23 MB/s | 3.0 MB 00:00 2026-03-09T19:49:01.101 INFO:teuthology.orchestra.run.vm02.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.5 MB/s | 45 kB 00:00 2026-03-09T19:49:01.121 INFO:teuthology.orchestra.run.vm02.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 18 MB/s | 662 kB 00:00 2026-03-09T19:49:01.134 INFO:teuthology.orchestra.run.vm02.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 7.5 MB/s | 246 kB 00:00 2026-03-09T19:49:01.154 INFO:teuthology.orchestra.run.vm02.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 7.1 MB/s | 233 kB 00:00 2026-03-09T19:49:01.167 INFO:teuthology.orchestra.run.vm02.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 8.8 MB/s | 292 kB 00:00 2026-03-09T19:49:01.200 INFO:teuthology.orchestra.run.vm02.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 4.0 MB/s | 188 kB 00:00 2026-03-09T19:49:01.221 INFO:teuthology.orchestra.run.vm02.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 781 kB/s | 42 kB 00:00 2026-03-09T19:49:01.310 INFO:teuthology.orchestra.run.vm02.stdout:(64/136): protobuf-3.14.0-17.el9.x86_64.rpm 11 MB/s | 1.0 MB 00:00 2026-03-09T19:49:01.370 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:49:01.371 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:49:01.424 INFO:teuthology.orchestra.run.vm02.stdout:(65/136): openblas-openmp-0.3.29-1.el9.x86_64.r 24 MB/s | 5.3 MB 00:00 2026-03-09T19:49:01.458 INFO:teuthology.orchestra.run.vm02.stdout:(66/136): python3-devel-3.9.25-3.el9.x86_64.rpm 7.2 MB/s | 244 kB 00:00 2026-03-09T19:49:01.491 INFO:teuthology.orchestra.run.vm02.stdout:(67/136): python3-jinja2-2.11.3-8.el9.noarch.rp 7.4 MB/s | 249 kB 00:00 2026-03-09T19:49:01.522 INFO:teuthology.orchestra.run.vm02.stdout:(68/136): python3-jmespath-1.0.1-1.el9.noarch.r 1.5 MB/s | 48 kB 00:00 2026-03-09T19:49:01.554 INFO:teuthology.orchestra.run.vm02.stdout:(69/136): python3-libstoragemgmt-1.10.1-1.el9.x 5.5 MB/s | 177 kB 00:00 2026-03-09T19:49:01.585 INFO:teuthology.orchestra.run.vm02.stdout:(70/136): python3-mako-1.1.4-6.el9.noarch.rpm 5.4 MB/s | 172 kB 00:00 2026-03-09T19:49:01.613 INFO:teuthology.orchestra.run.vm02.stdout:(71/136): python3-babel-2.9.1-2.el9.noarch.rpm 20 MB/s | 6.0 MB 00:00 2026-03-09T19:49:01.616 INFO:teuthology.orchestra.run.vm02.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 1.1 MB/s | 35 kB 00:00 2026-03-09T19:49:01.917 INFO:teuthology.orchestra.run.vm02.stdout:(73/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 20 MB/s | 6.1 MB 00:00 2026-03-09T19:49:01.918 INFO:teuthology.orchestra.run.vm02.stdout:(74/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.4 MB/s | 442 kB 00:00 2026-03-09T19:49:01.949 INFO:teuthology.orchestra.run.vm02.stdout:(75/136): python3-packaging-20.9-5.el9.noarch.r 2.3 MB/s | 77 kB 00:00 2026-03-09T19:49:01.951 INFO:teuthology.orchestra.run.vm11.stdout:CentOS Stream 9 - CRB 7.3 MB/s | 8.0 MB 00:01 2026-03-09T19:49:01.951 INFO:teuthology.orchestra.run.vm02.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 8.0 MB/s | 267 kB 00:00 2026-03-09T19:49:01.981 INFO:teuthology.orchestra.run.vm02.stdout:(77/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 5.0 MB/s | 157 kB 00:00 2026-03-09T19:49:01.984 INFO:teuthology.orchestra.run.vm02.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 8.3 MB/s | 277 kB 00:00 2026-03-09T19:49:02.011 INFO:teuthology.orchestra.run.vm02.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 1.8 MB/s | 54 kB 00:00 2026-03-09T19:49:02.040 INFO:teuthology.orchestra.run.vm02.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 1.4 MB/s | 42 kB 00:00 2026-03-09T19:49:02.083 INFO:teuthology.orchestra.run.vm02.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 5.5 MB/s | 240 kB 00:00 2026-03-09T19:49:02.114 INFO:teuthology.orchestra.run.vm02.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 1.2 MB/s | 37 kB 00:00 2026-03-09T19:49:02.148 INFO:teuthology.orchestra.run.vm02.stdout:(83/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 1.9 MB/s | 66 kB 00:00 2026-03-09T19:49:02.181 INFO:teuthology.orchestra.run.vm02.stdout:(84/136): socat-1.7.4.1-8.el9.x86_64.rpm 9.0 MB/s | 303 kB 00:00 2026-03-09T19:49:02.211 INFO:teuthology.orchestra.run.vm02.stdout:(85/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.1 MB/s | 64 kB 00:00 2026-03-09T19:49:02.309 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:49:02.325 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T19:49:02.338 INFO:teuthology.orchestra.run.vm10.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T19:49:02.519 INFO:teuthology.orchestra.run.vm10.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T19:49:02.521 INFO:teuthology.orchestra.run.vm10.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:02.544 INFO:teuthology.orchestra.run.vm02.stdout:(86/136): lua-devel-5.4.4-4.el9.x86_64.rpm 67 kB/s | 22 kB 00:00 2026-03-09T19:49:02.585 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:02.588 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:02.618 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:02.645 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:49:02.733 INFO:teuthology.orchestra.run.vm10.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T19:49:02.782 INFO:teuthology.orchestra.run.vm10.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T19:49:02.787 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T19:49:02.798 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T19:49:02.799 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:02.836 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:02.838 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:02.850 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:02.885 INFO:teuthology.orchestra.run.vm10.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T19:49:02.915 INFO:teuthology.orchestra.run.vm02.stdout:(87/136): protobuf-compiler-3.14.0-17.el9.x86_6 2.3 MB/s | 862 kB 00:00 2026-03-09T19:49:02.926 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T19:49:02.932 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T19:49:02.946 INFO:teuthology.orchestra.run.vm02.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 18 MB/s | 551 kB 00:00 2026-03-09T19:49:02.960 INFO:teuthology.orchestra.run.vm10.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T19:49:02.969 INFO:teuthology.orchestra.run.vm02.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 13 MB/s | 308 kB 00:00 2026-03-09T19:49:02.978 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T19:49:02.986 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T19:49:02.995 INFO:teuthology.orchestra.run.vm02.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 748 kB/s | 19 kB 00:00 2026-03-09T19:49:03.000 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T19:49:03.008 INFO:teuthology.orchestra.run.vm10.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T19:49:03.013 INFO:teuthology.orchestra.run.vm10.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T19:49:03.019 INFO:teuthology.orchestra.run.vm10.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T19:49:03.050 INFO:teuthology.orchestra.run.vm10.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T19:49:03.068 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T19:49:03.077 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T19:49:03.106 INFO:teuthology.orchestra.run.vm02.stdout:(91/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 17 MB/s | 19 MB 00:01 2026-03-09T19:49:03.111 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T19:49:03.113 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T19:49:03.125 INFO:teuthology.orchestra.run.vm02.stdout:(92/136): libarrow-9.0.0-15.el9.x86_64.rpm 34 MB/s | 4.4 MB 00:00 2026-03-09T19:49:03.128 INFO:teuthology.orchestra.run.vm02.stdout:(93/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 1.1 MB/s | 25 kB 00:00 2026-03-09T19:49:03.146 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T19:49:03.152 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T19:49:03.160 INFO:teuthology.orchestra.run.vm02.stdout:(94/136): liboath-2.6.12-1.el9.x86_64.rpm 1.4 MB/s | 49 kB 00:00 2026-03-09T19:49:03.162 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T19:49:03.163 INFO:teuthology.orchestra.run.vm02.stdout:(95/136): libunwind-1.6.2-1.el9.x86_64.rpm 1.9 MB/s | 67 kB 00:00 2026-03-09T19:49:03.177 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T19:49:03.185 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T19:49:03.190 INFO:teuthology.orchestra.run.vm02.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 30 MB/s | 838 kB 00:00 2026-03-09T19:49:03.203 INFO:teuthology.orchestra.run.vm02.stdout:(97/136): luarocks-3.9.2-5.el9.noarch.rpm 3.5 MB/s | 151 kB 00:00 2026-03-09T19:49:03.206 INFO:teuthology.orchestra.run.vm02.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 8.7 MB/s | 29 kB 00:00 2026-03-09T19:49:03.209 INFO:teuthology.orchestra.run.vm02.stdout:(99/136): python3-asyncssh-2.13.2-5.el9.noarch. 29 MB/s | 548 kB 00:00 2026-03-09T19:49:03.209 INFO:teuthology.orchestra.run.vm02.stdout:(100/136): python3-backports-tarfile-1.2.0-1.el 22 MB/s | 60 kB 00:00 2026-03-09T19:49:03.212 INFO:teuthology.orchestra.run.vm02.stdout:(101/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 14 MB/s | 43 kB 00:00 2026-03-09T19:49:03.213 INFO:teuthology.orchestra.run.vm02.stdout:(102/136): python3-cachetools-4.2.4-1.el9.noarc 10 MB/s | 32 kB 00:00 2026-03-09T19:49:03.215 INFO:teuthology.orchestra.run.vm02.stdout:(103/136): python3-certifi-2023.05.07-4.el9.noa 5.7 MB/s | 14 kB 00:00 2026-03-09T19:49:03.216 INFO:teuthology.orchestra.run.vm10.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T19:49:03.219 INFO:teuthology.orchestra.run.vm02.stdout:(104/136): python3-cheroot-10.0.1-4.el9.noarch. 26 MB/s | 173 kB 00:00 2026-03-09T19:49:03.222 INFO:teuthology.orchestra.run.vm10.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T19:49:03.223 INFO:teuthology.orchestra.run.vm02.stdout:(105/136): python3-cherrypy-18.6.1-2.el9.noarch 45 MB/s | 358 kB 00:00 2026-03-09T19:49:03.224 INFO:teuthology.orchestra.run.vm02.stdout:(106/136): python3-google-auth-2.45.0-1.el9.noa 54 MB/s | 254 kB 00:00 2026-03-09T19:49:03.230 INFO:teuthology.orchestra.run.vm02.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 26 MB/s | 144 kB 00:00 2026-03-09T19:49:03.230 INFO:teuthology.orchestra.run.vm10.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T19:49:03.235 INFO:teuthology.orchestra.run.vm02.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 2.5 MB/s | 11 kB 00:00 2026-03-09T19:49:03.240 INFO:teuthology.orchestra.run.vm02.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 3.6 MB/s | 18 kB 00:00 2026-03-09T19:49:03.247 INFO:teuthology.orchestra.run.vm02.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 3.6 MB/s | 23 kB 00:00 2026-03-09T19:49:03.253 INFO:teuthology.orchestra.run.vm02.stdout:(111/136): python3-grpcio-1.46.7-10.el9.x86_64. 68 MB/s | 2.0 MB 00:00 2026-03-09T19:49:03.254 INFO:teuthology.orchestra.run.vm02.stdout:(112/136): python3-jaraco-context-6.0.1-3.el9.n 2.8 MB/s | 20 kB 00:00 2026-03-09T19:49:03.255 INFO:teuthology.orchestra.run.vm02.stdout:(113/136): python3-jaraco-functools-3.5.0-2.el9 9.7 MB/s | 19 kB 00:00 2026-03-09T19:49:03.256 INFO:teuthology.orchestra.run.vm02.stdout:(114/136): python3-jaraco-text-4.0.0-2.el9.noar 13 MB/s | 26 kB 00:00 2026-03-09T19:49:03.262 INFO:teuthology.orchestra.run.vm02.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 8.5 MB/s | 46 kB 00:00 2026-03-09T19:49:03.262 INFO:teuthology.orchestra.run.vm10.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T19:49:03.269 INFO:teuthology.orchestra.run.vm02.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 11 MB/s | 79 kB 00:00 2026-03-09T19:49:03.272 INFO:teuthology.orchestra.run.vm02.stdout:(117/136): python3-kubernetes-26.1.0-3.el9.noar 63 MB/s | 1.0 MB 00:00 2026-03-09T19:49:03.272 INFO:teuthology.orchestra.run.vm02.stdout:(118/136): python3-natsort-7.1.1-5.el9.noarch.r 16 MB/s | 58 kB 00:00 2026-03-09T19:49:03.277 INFO:teuthology.orchestra.run.vm02.stdout:(119/136): python3-pecan-1.4.2-3.el9.noarch.rpm 55 MB/s | 272 kB 00:00 2026-03-09T19:49:03.277 INFO:teuthology.orchestra.run.vm02.stdout:(120/136): python3-portend-3.1.0-2.el9.noarch.r 3.3 MB/s | 16 kB 00:00 2026-03-09T19:49:03.280 INFO:teuthology.orchestra.run.vm02.stdout:(121/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 29 MB/s | 90 kB 00:00 2026-03-09T19:49:03.281 INFO:teuthology.orchestra.run.vm02.stdout:(122/136): python3-repoze-lru-0.7-16.el9.noarch 9.7 MB/s | 31 kB 00:00 2026-03-09T19:49:03.284 INFO:teuthology.orchestra.run.vm02.stdout:(123/136): python3-routes-2.5.1-5.el9.noarch.rp 48 MB/s | 188 kB 00:00 2026-03-09T19:49:03.286 INFO:teuthology.orchestra.run.vm02.stdout:(124/136): python3-rsa-4.9-2.el9.noarch.rpm 12 MB/s | 59 kB 00:00 2026-03-09T19:49:03.286 INFO:teuthology.orchestra.run.vm02.stdout:(125/136): python3-tempora-5.0.0-2.el9.noarch.r 17 MB/s | 36 kB 00:00 2026-03-09T19:49:03.289 INFO:teuthology.orchestra.run.vm02.stdout:(126/136): python3-typing-extensions-4.15.0-1.e 30 MB/s | 86 kB 00:00 2026-03-09T19:49:03.291 INFO:teuthology.orchestra.run.vm02.stdout:(127/136): python3-webob-1.8.8-2.el9.noarch.rpm 53 MB/s | 230 kB 00:00 2026-03-09T19:49:03.292 INFO:teuthology.orchestra.run.vm02.stdout:(128/136): python3-websocket-client-1.2.3-2.el9 31 MB/s | 90 kB 00:00 2026-03-09T19:49:03.297 INFO:teuthology.orchestra.run.vm02.stdout:(129/136): python3-werkzeug-2.0.3-3.el9.1.noarc 66 MB/s | 427 kB 00:00 2026-03-09T19:49:03.298 INFO:teuthology.orchestra.run.vm02.stdout:(130/136): python3-xmltodict-0.12.0-15.el9.noar 3.4 MB/s | 22 kB 00:00 2026-03-09T19:49:03.300 INFO:teuthology.orchestra.run.vm02.stdout:(131/136): python3-zc-lockfile-2.0-10.el9.noarc 7.2 MB/s | 20 kB 00:00 2026-03-09T19:49:03.302 INFO:teuthology.orchestra.run.vm02.stdout:(132/136): re2-20211101-20.el9.x86_64.rpm 48 MB/s | 191 kB 00:00 2026-03-09T19:49:03.321 INFO:teuthology.orchestra.run.vm02.stdout:(133/136): thrift-0.15.0-4.el9.x86_64.rpm 76 MB/s | 1.6 MB 00:00 2026-03-09T19:49:03.333 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T19:49:03.350 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T19:49:03.358 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T19:49:03.367 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T19:49:03.374 INFO:teuthology.orchestra.run.vm10.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T19:49:03.378 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T19:49:03.398 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T19:49:03.426 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T19:49:03.433 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T19:49:03.440 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T19:49:03.457 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T19:49:03.472 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T19:49:03.485 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T19:49:03.542 INFO:teuthology.orchestra.run.vm11.stdout:CentOS Stream 9 - Extras packages 27 kB/s | 20 kB 00:00 2026-03-09T19:49:03.554 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T19:49:03.563 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T19:49:03.574 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T19:49:03.628 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T19:49:03.701 INFO:teuthology.orchestra.run.vm02.stdout:(134/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 5.8 MB/s | 50 MB 00:08 2026-03-09T19:49:04.099 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T19:49:04.116 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T19:49:04.123 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T19:49:04.131 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T19:49:04.137 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T19:49:04.146 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T19:49:04.151 INFO:teuthology.orchestra.run.vm10.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T19:49:04.153 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T19:49:04.188 INFO:teuthology.orchestra.run.vm10.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T19:49:04.248 INFO:teuthology.orchestra.run.vm10.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T19:49:04.262 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T19:49:04.271 INFO:teuthology.orchestra.run.vm10.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T19:49:04.277 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T19:49:04.286 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T19:49:04.291 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T19:49:04.303 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T19:49:04.308 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T19:49:04.346 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T19:49:04.362 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T19:49:04.410 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T19:49:04.425 INFO:teuthology.orchestra.run.vm11.stdout:Extra Packages for Enterprise Linux 25 MB/s | 20 MB 00:00 2026-03-09T19:49:04.699 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T19:49:04.742 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T19:49:04.749 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T19:49:04.814 INFO:teuthology.orchestra.run.vm10.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T19:49:04.817 INFO:teuthology.orchestra.run.vm10.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T19:49:04.841 INFO:teuthology.orchestra.run.vm10.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T19:49:04.981 INFO:teuthology.orchestra.run.vm02.stdout:(135/136): librados2-19.2.3-678.ge911bdeb.el9.x 2.0 MB/s | 3.4 MB 00:01 2026-03-09T19:49:05.143 INFO:teuthology.orchestra.run.vm02.stdout:(136/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.7 MB/s | 3.2 MB 00:01 2026-03-09T19:49:05.146 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:49:05.146 INFO:teuthology.orchestra.run.vm02.stdout:Total 12 MB/s | 210 MB 00:18 2026-03-09T19:49:05.249 INFO:teuthology.orchestra.run.vm10.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T19:49:05.346 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T19:49:05.735 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:49:05.789 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:49:05.790 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:49:06.236 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T19:49:06.269 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T19:49:06.278 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T19:49:06.283 INFO:teuthology.orchestra.run.vm10.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T19:49:06.465 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T19:49:06.469 INFO:teuthology.orchestra.run.vm10.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:49:06.509 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:49:06.514 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T19:49:06.522 INFO:teuthology.orchestra.run.vm10.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T19:49:06.665 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:49:06.665 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:49:06.818 INFO:teuthology.orchestra.run.vm10.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T19:49:06.821 INFO:teuthology.orchestra.run.vm10.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:49:06.842 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:49:06.846 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T19:49:07.601 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:49:07.617 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T19:49:07.631 INFO:teuthology.orchestra.run.vm02.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T19:49:07.815 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T19:49:07.817 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:07.893 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:07.895 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:07.935 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:07.945 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:49:07.950 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T19:49:07.952 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T19:49:07.958 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T19:49:07.968 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T19:49:07.969 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:08.011 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:08.013 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:08.030 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:08.060 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:08.070 INFO:teuthology.orchestra.run.vm02.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T19:49:08.142 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:08.169 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T19:49:08.170 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:08.177 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T19:49:08.193 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T19:49:08.205 INFO:teuthology.orchestra.run.vm02.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T19:49:08.220 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T19:49:08.221 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T19:49:08.231 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T19:49:08.242 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T19:49:08.251 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T19:49:08.256 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T19:49:08.263 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T19:49:08.292 INFO:teuthology.orchestra.run.vm02.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T19:49:08.311 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T19:49:08.315 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T19:49:08.323 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T19:49:08.326 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T19:49:08.332 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T19:49:08.348 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T19:49:08.357 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T19:49:08.365 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T19:49:08.377 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T19:49:08.379 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T19:49:08.392 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T19:49:08.403 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T19:49:08.423 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T19:49:08.433 INFO:teuthology.orchestra.run.vm02.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T19:49:08.440 INFO:teuthology.orchestra.run.vm02.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T19:49:08.449 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T19:49:08.489 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T19:49:08.492 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T19:49:08.671 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T19:49:08.690 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T19:49:08.759 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T19:49:08.763 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:49:08.768 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T19:49:08.772 INFO:teuthology.orchestra.run.vm10.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T19:49:08.779 INFO:teuthology.orchestra.run.vm10.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T19:49:08.779 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T19:49:08.781 INFO:teuthology.orchestra.run.vm10.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:49:08.786 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T19:49:08.793 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T19:49:08.804 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:49:08.812 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T19:49:08.840 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T19:49:08.849 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T19:49:08.858 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T19:49:08.874 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T19:49:08.888 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T19:49:08.901 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T19:49:08.969 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T19:49:08.983 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T19:49:08.998 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T19:49:09.051 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T19:49:09.099 INFO:teuthology.orchestra.run.vm11.stdout:lab-extras 64 kB/s | 50 kB 00:00 2026-03-09T19:49:09.134 INFO:teuthology.orchestra.run.vm10.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T19:49:09.141 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:49:09.191 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:49:09.191 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T19:49:09.191 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T19:49:09.191 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:09.198 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:49:09.455 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T19:49:09.474 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T19:49:09.481 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T19:49:09.489 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T19:49:09.496 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T19:49:09.503 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T19:49:09.508 INFO:teuthology.orchestra.run.vm02.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T19:49:09.510 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T19:49:09.543 INFO:teuthology.orchestra.run.vm02.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T19:49:09.602 INFO:teuthology.orchestra.run.vm02.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T19:49:09.617 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T19:49:09.625 INFO:teuthology.orchestra.run.vm02.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T19:49:09.632 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T19:49:09.640 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T19:49:09.647 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T19:49:09.657 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T19:49:09.664 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T19:49:09.700 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T19:49:09.715 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T19:49:09.763 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T19:49:10.052 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T19:49:10.088 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T19:49:10.096 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T19:49:10.164 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T19:49:10.168 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T19:49:10.203 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T19:49:10.534 INFO:teuthology.orchestra.run.vm11.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:49:10.534 INFO:teuthology.orchestra.run.vm11.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-03-09T19:49:10.540 INFO:teuthology.orchestra.run.vm11.stdout:Package bzip2-1.0.8-11.el9.x86_64 is already installed. 2026-03-09T19:49:10.541 INFO:teuthology.orchestra.run.vm11.stdout:Package perl-Test-Harness-1:3.42-461.el9.noarch is already installed. 2026-03-09T19:49:10.570 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:49:10.574 INFO:teuthology.orchestra.run.vm11.stdout:====================================================================================== 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout:====================================================================================== 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout:Installing: 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 6.5 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.5 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.2 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 145 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.1 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 150 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 3.8 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 7.4 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 49 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 11 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 50 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 299 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 769 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 34 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 1.0 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 127 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 165 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 323 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 303 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 100 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 85 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.1 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 171 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout:Upgrading: 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.4 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 3.2 M 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout:Installing dependencies: 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-03-09T19:49:10.575 INFO:teuthology.orchestra.run.vm11.stdout: boost-program-options x86_64 1.75.0-13.el9 appstream 104 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 22 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 31 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 2.4 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 253 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 4.7 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 17 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 ceph-noarch 17 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 25 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: cryptsetup x86_64 2.8.1-3.el9 baseos 351 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas x86_64 3.0.4-9.el9 appstream 30 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 appstream 3.0 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 appstream 15 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 40 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 163 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libconfig x86_64 1.7.2-9.el9 baseos 72 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libgfortran x86_64 11.5.0-14.el9 baseos 794 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libnbd x86_64 1.20.3-4.el9 appstream 164 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 160 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libquadmath x86_64 11.5.0-14.el9 baseos 184 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 45 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 503 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 5.4 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 246 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: libxslt x86_64 1.1.34-12.el9 appstream 233 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 292 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: lua x86_64 5.4.4-4.el9 appstream 188 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: lua-devel x86_64 5.4.4-4.el9 crb 22 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: mailcap noarch 2.1.49-5.el9 baseos 33 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: openblas x86_64 0.3.29-1.el9 appstream 42 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-03-09T19:49:10.576 INFO:teuthology.orchestra.run.vm11.stdout: pciutils x86_64 3.7.0-7.el9 baseos 93 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: protobuf x86_64 3.14.0-17.el9 appstream 1.0 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-compiler x86_64 3.14.0-17.el9 crb 862 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-babel noarch 2.9.1-2.el9 appstream 6.0 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 45 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 ceph 142 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 253 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-cheroot noarch 10.0.1-4.el9 epel 173 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-cherrypy noarch 18.6.1-2.el9 epel 358 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-cryptography x86_64 36.0.1-5.el9 baseos 1.2 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-devel x86_64 3.9.25-3.el9 appstream 244 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-jinja2 noarch 2.11.3-8.el9 appstream 249 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 177 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-logutils noarch 0.3.5-21.el9 epel 46 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-mako noarch 1.1.4-6.el9 appstream 172 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 35 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy x86_64 1:1.23.5-2.el9 appstream 6.1 M 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 appstream 442 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-packaging noarch 20.9-5.el9 appstream 77 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-pecan noarch 1.4.2-3.el9 epel 272 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-ply noarch 3.11-14.el9 baseos 106 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-protobuf noarch 3.14.0-17.el9 appstream 267 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1 noarch 0.4.8-7.el9 appstream 157 k 2026-03-09T19:49:10.577 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 appstream 277 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-pycparser noarch 2.20-6.el9 baseos 135 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests noarch 2.25.1-10.el9 baseos 126 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 54 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-toml noarch 0.10.2-6.el9 appstream 42 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-webob noarch 1.8.8-2.el9 epel 230 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 epel 427 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: qatlib x86_64 25.08.0-2.el9 appstream 240 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 66 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: socat x86_64 1.7.4.1-8.el9 appstream 303 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: unzip x86_64 6.0-59.el9 baseos 182 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 64 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: zip x86_64 3.0-35.el9 baseos 266 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout:Installing weak dependencies: 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-service x86_64 25.08.0-2.el9 appstream 37 k 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout:====================================================================================== 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout:Install 134 Packages 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout:Upgrade 2 Packages 2026-03-09T19:49:10.578 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:49:10.579 INFO:teuthology.orchestra.run.vm11.stdout:Total download size: 210 M 2026-03-09T19:49:10.579 INFO:teuthology.orchestra.run.vm11.stdout:Downloading Packages: 2026-03-09T19:49:10.629 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T19:49:10.724 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T19:49:11.569 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T19:49:11.599 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T19:49:11.605 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T19:49:11.610 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T19:49:11.782 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T19:49:11.785 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:49:11.820 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:49:11.825 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T19:49:11.835 INFO:teuthology.orchestra.run.vm02.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T19:49:12.114 INFO:teuthology.orchestra.run.vm02.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T19:49:12.131 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:49:12.154 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:49:12.157 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T19:49:12.318 INFO:teuthology.orchestra.run.vm11.stdout:(1/136): ceph-19.2.3-678.ge911bdeb.el9.x86_64.r 13 kB/s | 6.5 kB 00:00 2026-03-09T19:49:13.283 INFO:teuthology.orchestra.run.vm11.stdout:(2/136): ceph-fuse-19.2.3-678.ge911bdeb.el9.x86 1.2 MB/s | 1.2 MB 00:00 2026-03-09T19:49:13.374 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:13.443 INFO:teuthology.orchestra.run.vm11.stdout:(3/136): ceph-immutable-object-cache-19.2.3-678 910 kB/s | 145 kB 00:00 2026-03-09T19:49:13.444 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:13.469 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:49:13.489 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T19:49:13.516 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T19:49:13.615 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T19:49:13.635 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T19:49:13.668 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T19:49:13.718 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T19:49:13.783 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T19:49:13.795 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T19:49:13.802 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:49:13.811 INFO:teuthology.orchestra.run.vm02.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T19:49:13.818 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T19:49:13.820 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:49:13.850 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:49:14.058 INFO:teuthology.orchestra.run.vm11.stdout:(4/136): ceph-base-19.2.3-678.ge911bdeb.el9.x86 2.5 MB/s | 5.5 MB 00:02 2026-03-09T19:49:14.188 INFO:teuthology.orchestra.run.vm11.stdout:(5/136): ceph-mds-19.2.3-678.ge911bdeb.el9.x86_ 3.3 MB/s | 2.4 MB 00:00 2026-03-09T19:49:14.206 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T19:49:14.213 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:49:14.262 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:49:14.262 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T19:49:14.262 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T19:49:14.262 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:14.280 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:49:14.299 INFO:teuthology.orchestra.run.vm11.stdout:(6/136): ceph-mgr-19.2.3-678.ge911bdeb.el9.x86_ 4.5 MB/s | 1.1 MB 00:00 2026-03-09T19:49:15.333 INFO:teuthology.orchestra.run.vm11.stdout:(7/136): ceph-mon-19.2.3-678.ge911bdeb.el9.x86_ 4.1 MB/s | 4.7 MB 00:01 2026-03-09T19:49:16.852 INFO:teuthology.orchestra.run.vm11.stdout:(8/136): ceph-osd-19.2.3-678.ge911bdeb.el9.x86_ 6.7 MB/s | 17 MB 00:02 2026-03-09T19:49:16.917 INFO:teuthology.orchestra.run.vm11.stdout:(9/136): ceph-common-19.2.3-678.ge911bdeb.el9.x 4.3 MB/s | 22 MB 00:05 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /sys 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /proc 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /mnt 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /var/tmp 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /home 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /root 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /tmp 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:16.970 INFO:teuthology.orchestra.run.vm11.stdout:(10/136): ceph-selinux-19.2.3-678.ge911bdeb.el9 213 kB/s | 25 kB 00:00 2026-03-09T19:49:17.087 INFO:teuthology.orchestra.run.vm11.stdout:(11/136): libcephfs-devel-19.2.3-678.ge911bdeb. 287 kB/s | 34 kB 00:00 2026-03-09T19:49:17.101 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:49:17.130 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:17.352 INFO:teuthology.orchestra.run.vm11.stdout:(12/136): ceph-radosgw-19.2.3-678.ge911bdeb.el9 5.3 MB/s | 11 MB 00:02 2026-03-09T19:49:17.358 INFO:teuthology.orchestra.run.vm11.stdout:(13/136): libcephfs2-19.2.3-678.ge911bdeb.el9.x 3.6 MB/s | 1.0 MB 00:00 2026-03-09T19:49:17.385 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:49:17.415 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:17.428 INFO:teuthology.orchestra.run.vm10.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T19:49:17.431 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T19:49:17.454 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:17.454 INFO:teuthology.orchestra.run.vm10.stdout:Creating group 'qat' with GID 994. 2026-03-09T19:49:17.454 INFO:teuthology.orchestra.run.vm10.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T19:49:17.454 INFO:teuthology.orchestra.run.vm10.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T19:49:17.454 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:17.467 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:17.477 INFO:teuthology.orchestra.run.vm11.stdout:(14/136): libcephsqlite-19.2.3-678.ge911bdeb.el 1.3 MB/s | 163 kB 00:00 2026-03-09T19:49:17.479 INFO:teuthology.orchestra.run.vm11.stdout:(15/136): librados-devel-19.2.3-678.ge911bdeb.e 1.0 MB/s | 127 kB 00:00 2026-03-09T19:49:17.498 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:17.498 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T19:49:17.498 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:17.541 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T19:49:17.622 INFO:teuthology.orchestra.run.vm10.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T19:49:17.627 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:49:17.642 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:49:17.642 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:17.642 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:49:17.642 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:18.501 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:49:18.527 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:49:18.528 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:18.528 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:49:18.528 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:49:18.528 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:49:18.528 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:18.586 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:49:18.589 INFO:teuthology.orchestra.run.vm10.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:49:18.596 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T19:49:18.622 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T19:49:18.626 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:49:19.002 INFO:teuthology.orchestra.run.vm11.stdout:(16/136): libradosstriper1-19.2.3-678.ge911bdeb 330 kB/s | 503 kB 00:01 2026-03-09T19:49:19.125 INFO:teuthology.orchestra.run.vm11.stdout:(17/136): python3-ceph-argparse-19.2.3-678.ge91 367 kB/s | 45 kB 00:00 2026-03-09T19:49:19.237 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:49:19.245 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:49:19.248 INFO:teuthology.orchestra.run.vm11.stdout:(18/136): python3-ceph-common-19.2.3-678.ge911b 1.1 MB/s | 142 kB 00:00 2026-03-09T19:49:19.372 INFO:teuthology.orchestra.run.vm11.stdout:(19/136): python3-cephfs-19.2.3-678.ge911bdeb.e 1.3 MB/s | 165 kB 00:00 2026-03-09T19:49:19.404 INFO:teuthology.orchestra.run.vm11.stdout:(20/136): librgw2-19.2.3-678.ge911bdeb.el9.x86_ 2.8 MB/s | 5.4 MB 00:01 2026-03-09T19:49:19.555 INFO:teuthology.orchestra.run.vm11.stdout:(21/136): python3-rados-19.2.3-678.ge911bdeb.el 1.7 MB/s | 323 kB 00:00 2026-03-09T19:49:19.675 INFO:teuthology.orchestra.run.vm11.stdout:(22/136): python3-rbd-19.2.3-678.ge911bdeb.el9. 1.1 MB/s | 303 kB 00:00 2026-03-09T19:49:19.678 INFO:teuthology.orchestra.run.vm11.stdout:(23/136): python3-rgw-19.2.3-678.ge911bdeb.el9. 811 kB/s | 100 kB 00:00 2026-03-09T19:49:19.793 INFO:teuthology.orchestra.run.vm11.stdout:(24/136): rbd-fuse-19.2.3-678.ge911bdeb.el9.x86 718 kB/s | 85 kB 00:00 2026-03-09T19:49:19.870 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:49:19.875 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:49:19.913 INFO:teuthology.orchestra.run.vm11.stdout:(25/136): rbd-nbd-19.2.3-678.ge911bdeb.el9.x86_ 1.4 MB/s | 171 kB 00:00 2026-03-09T19:49:19.960 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:49:20.027 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T19:49:20.047 INFO:teuthology.orchestra.run.vm11.stdout:(26/136): ceph-grafana-dashboards-19.2.3-678.ge 233 kB/s | 31 kB 00:00 2026-03-09T19:49:20.141 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:49:20.168 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:49:20.179 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:20.179 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:49:20.179 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:49:20.179 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:49:20.179 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:20.263 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:49:20.277 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:49:20.375 INFO:teuthology.orchestra.run.vm11.stdout:(27/136): ceph-mgr-cephadm-19.2.3-678.ge911bdeb 458 kB/s | 150 kB 00:00 2026-03-09T19:49:20.575 INFO:teuthology.orchestra.run.vm11.stdout:(28/136): rbd-mirror-19.2.3-678.ge911bdeb.el9.x 3.5 MB/s | 3.1 MB 00:00 2026-03-09T19:49:20.962 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T19:49:21.046 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:49:21.069 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:49:21.070 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:21.070 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:49:21.070 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:49:21.070 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:49:21.070 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:21.096 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:49:21.122 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:49:21.122 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:21.122 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:49:21.122 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:21.293 INFO:teuthology.orchestra.run.vm10.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:49:21.321 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:49:21.322 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:21.322 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:49:21.322 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:49:21.322 INFO:teuthology.orchestra.run.vm10.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:49:21.322 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:21.685 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-09T19:49:21.707 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:21.823 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:49:21.854 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:49:21.854 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:21.854 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:49:21.854 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:49:21.855 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:49:21.855 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:22.141 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:49:22.168 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:22.187 INFO:teuthology.orchestra.run.vm02.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T19:49:22.191 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T19:49:22.211 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:22.212 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'qat' with GID 994. 2026-03-09T19:49:22.212 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T19:49:22.212 INFO:teuthology.orchestra.run.vm02.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T19:49:22.212 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:22.223 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:22.255 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:49:22.255 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T19:49:22.255 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:22.301 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T19:49:22.389 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T19:49:22.394 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:49:22.410 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:49:22.410 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:22.410 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:49:22.410 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:22.616 INFO:teuthology.orchestra.run.vm11.stdout:(29/136): ceph-mgr-dashboard-19.2.3-678.ge911bd 1.7 MB/s | 3.8 MB 00:02 2026-03-09T19:49:22.953 INFO:teuthology.orchestra.run.vm11.stdout:(30/136): ceph-mgr-modules-core-19.2.3-678.ge91 751 kB/s | 253 kB 00:00 2026-03-09T19:49:23.057 INFO:teuthology.orchestra.run.vm11.stdout:(31/136): ceph-mgr-diskprediction-local-19.2.3- 3.0 MB/s | 7.4 MB 00:02 2026-03-09T19:49:23.072 INFO:teuthology.orchestra.run.vm11.stdout:(32/136): ceph-mgr-rook-19.2.3-678.ge911bdeb.el 418 kB/s | 49 kB 00:00 2026-03-09T19:49:23.199 INFO:teuthology.orchestra.run.vm11.stdout:(33/136): ceph-prometheus-alerts-19.2.3-678.ge9 118 kB/s | 17 kB 00:00 2026-03-09T19:49:23.275 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:49:23.308 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:23.319 INFO:teuthology.orchestra.run.vm11.stdout:(34/136): ceph-volume-19.2.3-678.ge911bdeb.el9. 1.2 MB/s | 299 kB 00:00 2026-03-09T19:49:23.378 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:49:23.382 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:49:23.389 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T19:49:23.438 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T19:49:23.442 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:49:23.684 INFO:teuthology.orchestra.run.vm11.stdout:(35/136): cephadm-19.2.3-678.ge911bdeb.el9.noar 1.6 MB/s | 769 kB 00:00 2026-03-09T19:49:23.886 INFO:teuthology.orchestra.run.vm11.stdout:(36/136): cryptsetup-2.8.1-3.el9.x86_64.rpm 619 kB/s | 351 kB 00:00 2026-03-09T19:49:24.027 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:49:24.028 INFO:teuthology.orchestra.run.vm11.stdout:(37/136): ledmon-libs-1.1.0-3.el9.x86_64.rpm 118 kB/s | 40 kB 00:00 2026-03-09T19:49:24.034 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:49:24.078 INFO:teuthology.orchestra.run.vm11.stdout:(38/136): libconfig-1.7.2-9.el9.x86_64.rpm 377 kB/s | 72 kB 00:00 2026-03-09T19:49:24.114 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T19:49:24.126 INFO:teuthology.orchestra.run.vm10.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T19:49:24.133 INFO:teuthology.orchestra.run.vm10.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T19:49:24.211 INFO:teuthology.orchestra.run.vm10.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T19:49:24.223 INFO:teuthology.orchestra.run.vm10.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:49:24.229 INFO:teuthology.orchestra.run.vm10.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T19:49:24.229 INFO:teuthology.orchestra.run.vm10.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:49:24.250 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:49:24.251 INFO:teuthology.orchestra.run.vm10.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:24.434 INFO:teuthology.orchestra.run.vm11.stdout:(39/136): libquadmath-11.5.0-14.el9.x86_64.rpm 519 kB/s | 184 kB 00:00 2026-03-09T19:49:24.588 INFO:teuthology.orchestra.run.vm11.stdout:(40/136): mailcap-2.1.49-5.el9.noarch.rpm 216 kB/s | 33 kB 00:00 2026-03-09T19:49:24.628 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:49:24.631 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:49:24.696 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:49:24.757 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T19:49:24.792 INFO:teuthology.orchestra.run.vm11.stdout:(41/136): libgfortran-11.5.0-14.el9.x86_64.rpm 1.0 MB/s | 794 kB 00:00 2026-03-09T19:49:24.792 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:49:24.819 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:24.836 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:49:24.849 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:49:24.906 INFO:teuthology.orchestra.run.vm11.stdout:(42/136): pciutils-3.7.0-7.el9.x86_64.rpm 293 kB/s | 93 kB 00:00 2026-03-09T19:49:25.026 INFO:teuthology.orchestra.run.vm11.stdout:(43/136): python3-cffi-1.14.5-5.el9.x86_64.rpm 1.1 MB/s | 253 kB 00:00 2026-03-09T19:49:25.240 INFO:teuthology.orchestra.run.vm11.stdout:(44/136): python3-ply-3.11-14.el9.noarch.rpm 497 kB/s | 106 kB 00:00 2026-03-09T19:49:25.249 INFO:teuthology.orchestra.run.vm11.stdout:(45/136): python3-cryptography-36.0.1-5.el9.x86 3.6 MB/s | 1.2 MB 00:00 2026-03-09T19:49:25.407 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T19:49:25.450 INFO:teuthology.orchestra.run.vm11.stdout:(46/136): python3-pycparser-2.20-6.el9.noarch.r 647 kB/s | 135 kB 00:00 2026-03-09T19:49:25.477 INFO:teuthology.orchestra.run.vm11.stdout:(47/136): python3-requests-2.25.1-10.el9.noarch 554 kB/s | 126 kB 00:00 2026-03-09T19:49:25.488 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:49:25.512 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:49:25.512 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:25.513 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:49:25.513 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:49:25.513 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:49:25.513 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:25.596 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:49:25.621 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:49:25.621 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:25.621 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:49:25.621 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:25.640 INFO:teuthology.orchestra.run.vm11.stdout:(48/136): python3-urllib3-1.26.5-7.el9.noarch.r 1.1 MB/s | 218 kB 00:00 2026-03-09T19:49:25.666 INFO:teuthology.orchestra.run.vm11.stdout:(49/136): unzip-6.0-59.el9.x86_64.rpm 965 kB/s | 182 kB 00:00 2026-03-09T19:49:25.823 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:49:25.851 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:25.865 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T19:49:25.866 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T19:49:25.867 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T19:49:25.868 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T19:49:25.869 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T19:49:25.870 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:49:25.871 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T19:49:25.872 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T19:49:25.873 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T19:49:25.890 INFO:teuthology.orchestra.run.vm11.stdout:(50/136): zip-3.0-35.el9.x86_64.rpm 1.0 MB/s | 266 kB 00:00 2026-03-09T19:49:25.915 INFO:teuthology.orchestra.run.vm11.stdout:(51/136): boost-program-options-1.75.0-13.el9.x 418 kB/s | 104 kB 00:00 2026-03-09T19:49:25.929 INFO:teuthology.orchestra.run.vm11.stdout:(52/136): flexiblas-3.0.4-9.el9.x86_64.rpm 762 kB/s | 30 kB 00:00 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout:Upgraded: 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout:Installed: 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.986 INFO:teuthology.orchestra.run.vm10.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:49:25.987 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:49:25.988 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:49:25.989 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:49:26.012 INFO:teuthology.orchestra.run.vm11.stdout:(53/136): flexiblas-openblas-openmp-3.0.4-9.el9 180 kB/s | 15 kB 00:00 2026-03-09T19:49:26.086 DEBUG:teuthology.parallel:result is None 2026-03-09T19:49:26.121 INFO:teuthology.orchestra.run.vm11.stdout:(54/136): flexiblas-netlib-3.0.4-9.el9.x86_64.r 15 MB/s | 3.0 MB 00:00 2026-03-09T19:49:26.122 INFO:teuthology.orchestra.run.vm11.stdout:(55/136): libnbd-1.20.3-4.el9.x86_64.rpm 1.5 MB/s | 164 kB 00:00 2026-03-09T19:49:26.157 INFO:teuthology.orchestra.run.vm11.stdout:(56/136): libpmemobj-1.12.1-1.el9.x86_64.rpm 4.6 MB/s | 160 kB 00:00 2026-03-09T19:49:26.157 INFO:teuthology.orchestra.run.vm11.stdout:(57/136): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.3 MB/s | 45 kB 00:00 2026-03-09T19:49:26.228 INFO:teuthology.orchestra.run.vm11.stdout:(58/136): librdkafka-1.6.1-102.el9.x86_64.rpm 9.1 MB/s | 662 kB 00:00 2026-03-09T19:49:26.229 INFO:teuthology.orchestra.run.vm11.stdout:(59/136): libstoragemgmt-1.10.1-1.el9.x86_64.rp 3.3 MB/s | 246 kB 00:00 2026-03-09T19:49:26.259 INFO:teuthology.orchestra.run.vm11.stdout:(60/136): libxslt-1.1.34-12.el9.x86_64.rpm 7.5 MB/s | 233 kB 00:00 2026-03-09T19:49:26.306 INFO:teuthology.orchestra.run.vm11.stdout:(61/136): lttng-ust-2.12.0-6.el9.x86_64.rpm 3.8 MB/s | 292 kB 00:00 2026-03-09T19:49:26.308 INFO:teuthology.orchestra.run.vm11.stdout:(62/136): lua-5.4.4-4.el9.x86_64.rpm 3.8 MB/s | 188 kB 00:00 2026-03-09T19:49:26.342 INFO:teuthology.orchestra.run.vm11.stdout:(63/136): openblas-0.3.29-1.el9.x86_64.rpm 1.2 MB/s | 42 kB 00:00 2026-03-09T19:49:26.655 INFO:teuthology.orchestra.run.vm11.stdout:(64/136): openblas-openmp-0.3.29-1.el9.x86_64.r 15 MB/s | 5.3 MB 00:00 2026-03-09T19:49:26.694 INFO:teuthology.orchestra.run.vm11.stdout:(65/136): protobuf-3.14.0-17.el9.x86_64.rpm 2.9 MB/s | 1.0 MB 00:00 2026-03-09T19:49:27.021 INFO:teuthology.orchestra.run.vm11.stdout:(66/136): python3-babel-2.9.1-2.el9.noarch.rpm 16 MB/s | 6.0 MB 00:00 2026-03-09T19:49:27.022 INFO:teuthology.orchestra.run.vm11.stdout:(67/136): python3-devel-3.9.25-3.el9.x86_64.rpm 746 kB/s | 244 kB 00:00 2026-03-09T19:49:27.065 INFO:teuthology.orchestra.run.vm11.stdout:(68/136): python3-jinja2-2.11.3-8.el9.noarch.rp 5.6 MB/s | 249 kB 00:00 2026-03-09T19:49:27.096 INFO:teuthology.orchestra.run.vm11.stdout:(69/136): python3-jmespath-1.0.1-1.el9.noarch.r 651 kB/s | 48 kB 00:00 2026-03-09T19:49:27.120 INFO:teuthology.orchestra.run.vm11.stdout:(70/136): python3-libstoragemgmt-1.10.1-1.el9.x 3.2 MB/s | 177 kB 00:00 2026-03-09T19:49:27.142 INFO:teuthology.orchestra.run.vm11.stdout:(71/136): python3-mako-1.1.4-6.el9.noarch.rpm 3.7 MB/s | 172 kB 00:00 2026-03-09T19:49:27.158 INFO:teuthology.orchestra.run.vm11.stdout:(72/136): python3-markupsafe-1.1.1-12.el9.x86_6 898 kB/s | 35 kB 00:00 2026-03-09T19:49:27.497 INFO:teuthology.orchestra.run.vm11.stdout:(73/136): python3-numpy-1.23.5-2.el9.x86_64.rpm 17 MB/s | 6.1 MB 00:00 2026-03-09T19:49:27.507 INFO:teuthology.orchestra.run.vm11.stdout:(74/136): python3-numpy-f2py-1.23.5-2.el9.x86_6 1.2 MB/s | 442 kB 00:00 2026-03-09T19:49:27.545 INFO:teuthology.orchestra.run.vm11.stdout:(75/136): python3-packaging-20.9-5.el9.noarch.r 1.6 MB/s | 77 kB 00:00 2026-03-09T19:49:27.559 INFO:teuthology.orchestra.run.vm11.stdout:(76/136): python3-protobuf-3.14.0-17.el9.noarch 5.0 MB/s | 267 kB 00:00 2026-03-09T19:49:27.589 INFO:teuthology.orchestra.run.vm11.stdout:(77/136): python3-pyasn1-0.4.8-7.el9.noarch.rpm 3.5 MB/s | 157 kB 00:00 2026-03-09T19:49:27.618 INFO:teuthology.orchestra.run.vm11.stdout:(78/136): python3-pyasn1-modules-0.4.8-7.el9.no 4.6 MB/s | 277 kB 00:00 2026-03-09T19:49:27.629 INFO:teuthology.orchestra.run.vm11.stdout:(79/136): python3-requests-oauthlib-1.3.0-12.el 1.3 MB/s | 54 kB 00:00 2026-03-09T19:49:27.910 INFO:teuthology.orchestra.run.vm11.stdout:(80/136): python3-toml-0.10.2-6.el9.noarch.rpm 149 kB/s | 42 kB 00:00 2026-03-09T19:49:28.293 INFO:teuthology.orchestra.run.vm11.stdout:(81/136): qatlib-25.08.0-2.el9.x86_64.rpm 627 kB/s | 240 kB 00:00 2026-03-09T19:49:28.452 INFO:teuthology.orchestra.run.vm11.stdout:(82/136): qatlib-service-25.08.0-2.el9.x86_64.r 233 kB/s | 37 kB 00:00 2026-03-09T19:49:28.652 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T19:49:28.678 INFO:teuthology.orchestra.run.vm11.stdout:(83/136): python3-scipy-1.9.3-2.el9.x86_64.rpm 18 MB/s | 19 MB 00:01 2026-03-09T19:49:28.679 INFO:teuthology.orchestra.run.vm11.stdout:(84/136): qatzip-libs-1.3.1-1.el9.x86_64.rpm 293 kB/s | 66 kB 00:00 2026-03-09T19:49:28.689 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T19:49:28.694 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T19:49:28.734 INFO:teuthology.orchestra.run.vm11.stdout:(85/136): socat-1.7.4.1-8.el9.x86_64.rpm 5.3 MB/s | 303 kB 00:00 2026-03-09T19:49:28.735 INFO:teuthology.orchestra.run.vm11.stdout:(86/136): xmlstarlet-1.6.1-20.el9.x86_64.rpm 1.1 MB/s | 64 kB 00:00 2026-03-09T19:49:28.755 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T19:49:28.766 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:49:28.771 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T19:49:28.771 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:49:28.791 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:49:28.791 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:28.971 INFO:teuthology.orchestra.run.vm11.stdout:(87/136): lua-devel-5.4.4-4.el9.x86_64.rpm 94 kB/s | 22 kB 00:00 2026-03-09T19:49:28.986 INFO:teuthology.orchestra.run.vm11.stdout:(88/136): abseil-cpp-20211102.0-4.el9.x86_64.rp 39 MB/s | 551 kB 00:00 2026-03-09T19:49:28.992 INFO:teuthology.orchestra.run.vm11.stdout:(89/136): gperftools-libs-2.9.1-3.el9.x86_64.rp 51 MB/s | 308 kB 00:00 2026-03-09T19:49:28.994 INFO:teuthology.orchestra.run.vm11.stdout:(90/136): grpc-data-1.46.7-10.el9.noarch.rpm 9.2 MB/s | 19 kB 00:00 2026-03-09T19:49:29.069 INFO:teuthology.orchestra.run.vm11.stdout:(91/136): libarrow-9.0.0-15.el9.x86_64.rpm 59 MB/s | 4.4 MB 00:00 2026-03-09T19:49:29.101 INFO:teuthology.orchestra.run.vm11.stdout:(92/136): libarrow-doc-9.0.0-15.el9.noarch.rpm 789 kB/s | 25 kB 00:00 2026-03-09T19:49:29.111 INFO:teuthology.orchestra.run.vm11.stdout:(93/136): liboath-2.6.12-1.el9.x86_64.rpm 4.9 MB/s | 49 kB 00:00 2026-03-09T19:49:29.114 INFO:teuthology.orchestra.run.vm11.stdout:(94/136): libunwind-1.6.2-1.el9.x86_64.rpm 26 MB/s | 67 kB 00:00 2026-03-09T19:49:29.118 INFO:teuthology.orchestra.run.vm11.stdout:(95/136): luarocks-3.9.2-5.el9.noarch.rpm 36 MB/s | 151 kB 00:00 2026-03-09T19:49:29.130 INFO:teuthology.orchestra.run.vm11.stdout:(96/136): parquet-libs-9.0.0-15.el9.x86_64.rpm 69 MB/s | 838 kB 00:00 2026-03-09T19:49:29.138 INFO:teuthology.orchestra.run.vm11.stdout:(97/136): python3-asyncssh-2.13.2-5.el9.noarch. 67 MB/s | 548 kB 00:00 2026-03-09T19:49:29.141 INFO:teuthology.orchestra.run.vm11.stdout:(98/136): python3-autocommand-2.2.2-8.el9.noarc 13 MB/s | 29 kB 00:00 2026-03-09T19:49:29.143 INFO:teuthology.orchestra.run.vm11.stdout:(99/136): python3-backports-tarfile-1.2.0-1.el9 24 MB/s | 60 kB 00:00 2026-03-09T19:49:29.145 INFO:teuthology.orchestra.run.vm11.stdout:(100/136): python3-bcrypt-3.2.2-1.el9.x86_64.rp 20 MB/s | 43 kB 00:00 2026-03-09T19:49:29.148 INFO:teuthology.orchestra.run.vm11.stdout:(101/136): python3-cachetools-4.2.4-1.el9.noarc 11 MB/s | 32 kB 00:00 2026-03-09T19:49:29.151 INFO:teuthology.orchestra.run.vm11.stdout:(102/136): python3-certifi-2023.05.07-4.el9.noa 7.0 MB/s | 14 kB 00:00 2026-03-09T19:49:29.155 INFO:teuthology.orchestra.run.vm11.stdout:(103/136): python3-cheroot-10.0.1-4.el9.noarch. 36 MB/s | 173 kB 00:00 2026-03-09T19:49:29.162 INFO:teuthology.orchestra.run.vm11.stdout:(104/136): python3-cherrypy-18.6.1-2.el9.noarch 60 MB/s | 358 kB 00:00 2026-03-09T19:49:29.166 INFO:teuthology.orchestra.run.vm11.stdout:(105/136): python3-google-auth-2.45.0-1.el9.noa 56 MB/s | 254 kB 00:00 2026-03-09T19:49:29.191 INFO:teuthology.orchestra.run.vm11.stdout:(106/136): python3-grpcio-1.46.7-10.el9.x86_64. 84 MB/s | 2.0 MB 00:00 2026-03-09T19:49:29.195 INFO:teuthology.orchestra.run.vm11.stdout:(107/136): python3-grpcio-tools-1.46.7-10.el9.x 38 MB/s | 144 kB 00:00 2026-03-09T19:49:29.197 INFO:teuthology.orchestra.run.vm11.stdout:(108/136): python3-jaraco-8.2.1-3.el9.noarch.rp 4.6 MB/s | 11 kB 00:00 2026-03-09T19:49:29.200 INFO:teuthology.orchestra.run.vm11.stdout:(109/136): python3-jaraco-classes-3.2.1-5.el9.n 7.6 MB/s | 18 kB 00:00 2026-03-09T19:49:29.202 INFO:teuthology.orchestra.run.vm11.stdout:(110/136): python3-jaraco-collections-3.0.0-8.e 11 MB/s | 23 kB 00:00 2026-03-09T19:49:29.204 INFO:teuthology.orchestra.run.vm11.stdout:(111/136): python3-jaraco-context-6.0.1-3.el9.n 9.7 MB/s | 20 kB 00:00 2026-03-09T19:49:29.207 INFO:teuthology.orchestra.run.vm11.stdout:(112/136): python3-jaraco-functools-3.5.0-2.el9 6.3 MB/s | 19 kB 00:00 2026-03-09T19:49:29.210 INFO:teuthology.orchestra.run.vm11.stdout:(113/136): python3-jaraco-text-4.0.0-2.el9.noar 9.9 MB/s | 26 kB 00:00 2026-03-09T19:49:29.225 INFO:teuthology.orchestra.run.vm11.stdout:(114/136): python3-kubernetes-26.1.0-3.el9.noar 72 MB/s | 1.0 MB 00:00 2026-03-09T19:49:29.227 INFO:teuthology.orchestra.run.vm11.stdout:(115/136): python3-logutils-0.3.5-21.el9.noarch 19 MB/s | 46 kB 00:00 2026-03-09T19:49:29.230 INFO:teuthology.orchestra.run.vm11.stdout:(116/136): python3-more-itertools-8.12.0-2.el9. 26 MB/s | 79 kB 00:00 2026-03-09T19:49:29.234 INFO:teuthology.orchestra.run.vm11.stdout:(117/136): python3-natsort-7.1.1-5.el9.noarch.r 17 MB/s | 58 kB 00:00 2026-03-09T19:49:29.239 INFO:teuthology.orchestra.run.vm11.stdout:(118/136): python3-pecan-1.4.2-3.el9.noarch.rpm 55 MB/s | 272 kB 00:00 2026-03-09T19:49:29.243 INFO:teuthology.orchestra.run.vm11.stdout:(119/136): python3-portend-3.1.0-2.el9.noarch.r 4.1 MB/s | 16 kB 00:00 2026-03-09T19:49:29.247 INFO:teuthology.orchestra.run.vm11.stdout:(120/136): python3-pyOpenSSL-21.0.0-1.el9.noarc 24 MB/s | 90 kB 00:00 2026-03-09T19:49:29.249 INFO:teuthology.orchestra.run.vm11.stdout:(121/136): python3-repoze-lru-0.7-16.el9.noarch 15 MB/s | 31 kB 00:00 2026-03-09T19:49:29.254 INFO:teuthology.orchestra.run.vm11.stdout:(122/136): python3-routes-2.5.1-5.el9.noarch.rp 40 MB/s | 188 kB 00:00 2026-03-09T19:49:29.257 INFO:teuthology.orchestra.run.vm11.stdout:(123/136): python3-rsa-4.9-2.el9.noarch.rpm 20 MB/s | 59 kB 00:00 2026-03-09T19:49:29.260 INFO:teuthology.orchestra.run.vm11.stdout:(124/136): python3-tempora-5.0.0-2.el9.noarch.r 14 MB/s | 36 kB 00:00 2026-03-09T19:49:29.263 INFO:teuthology.orchestra.run.vm11.stdout:(125/136): python3-typing-extensions-4.15.0-1.e 33 MB/s | 86 kB 00:00 2026-03-09T19:49:29.267 INFO:teuthology.orchestra.run.vm11.stdout:(126/136): python3-webob-1.8.8-2.el9.noarch.rpm 51 MB/s | 230 kB 00:00 2026-03-09T19:49:29.270 INFO:teuthology.orchestra.run.vm11.stdout:(127/136): python3-websocket-client-1.2.3-2.el9 32 MB/s | 90 kB 00:00 2026-03-09T19:49:29.277 INFO:teuthology.orchestra.run.vm11.stdout:(128/136): python3-werkzeug-2.0.3-3.el9.1.noarc 65 MB/s | 427 kB 00:00 2026-03-09T19:49:29.280 INFO:teuthology.orchestra.run.vm11.stdout:(129/136): python3-xmltodict-0.12.0-15.el9.noar 7.3 MB/s | 22 kB 00:00 2026-03-09T19:49:29.282 INFO:teuthology.orchestra.run.vm11.stdout:(130/136): python3-zc-lockfile-2.0-10.el9.noarc 10 MB/s | 20 kB 00:00 2026-03-09T19:49:29.286 INFO:teuthology.orchestra.run.vm11.stdout:(131/136): re2-20211101-20.el9.x86_64.rpm 46 MB/s | 191 kB 00:00 2026-03-09T19:49:29.306 INFO:teuthology.orchestra.run.vm11.stdout:(132/136): thrift-0.15.0-4.el9.x86_64.rpm 81 MB/s | 1.6 MB 00:00 2026-03-09T19:49:29.848 INFO:teuthology.orchestra.run.vm11.stdout:(133/136): protobuf-compiler-3.14.0-17.el9.x86_ 775 kB/s | 862 kB 00:01 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T19:49:30.419 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T19:49:30.420 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T19:49:30.422 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T19:49:30.423 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout:Upgraded: 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:49:30.538 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:49:30.539 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:49:30.540 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:49:30.633 DEBUG:teuthology.parallel:result is None 2026-03-09T19:49:31.474 INFO:teuthology.orchestra.run.vm11.stdout:(134/136): librados2-19.2.3-678.ge911bdeb.el9.x 1.6 MB/s | 3.4 MB 00:02 2026-03-09T19:49:32.489 INFO:teuthology.orchestra.run.vm11.stdout:(135/136): librbd1-19.2.3-678.ge911bdeb.el9.x86 1.2 MB/s | 3.2 MB 00:02 2026-03-09T19:49:55.059 INFO:teuthology.orchestra.run.vm11.stdout:(136/136): ceph-test-19.2.3-678.ge911bdeb.el9.x 1.3 MB/s | 50 MB 00:38 2026-03-09T19:49:55.063 INFO:teuthology.orchestra.run.vm11.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:49:55.063 INFO:teuthology.orchestra.run.vm11.stdout:Total 4.7 MB/s | 210 MB 00:44 2026-03-09T19:49:55.674 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:49:55.729 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:49:55.730 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:49:56.584 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:49:56.584 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:49:57.485 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:49:57.501 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 1/138 2026-03-09T19:49:57.513 INFO:teuthology.orchestra.run.vm11.stdout: Installing : thrift-0.15.0-4.el9.x86_64 2/138 2026-03-09T19:49:57.687 INFO:teuthology.orchestra.run.vm11.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/138 2026-03-09T19:49:57.689 INFO:teuthology.orchestra.run.vm11.stdout: Upgrading : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:57.753 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:49:57.754 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:57.788 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 5/138 2026-03-09T19:49:57.796 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:49:57.800 INFO:teuthology.orchestra.run.vm11.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 7/138 2026-03-09T19:49:57.803 INFO:teuthology.orchestra.run.vm11.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 8/138 2026-03-09T19:49:57.808 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 9/138 2026-03-09T19:49:57.818 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 10/138 2026-03-09T19:49:57.819 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:57.858 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:49:57.860 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:57.874 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 12/138 2026-03-09T19:49:57.906 INFO:teuthology.orchestra.run.vm11.stdout: Installing : re2-1:20211101-20.el9.x86_64 13/138 2026-03-09T19:49:57.943 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 14/138 2026-03-09T19:49:57.947 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-werkzeug-2.0.3-3.el9.1.noarch 15/138 2026-03-09T19:49:57.972 INFO:teuthology.orchestra.run.vm11.stdout: Installing : liboath-2.6.12-1.el9.x86_64 16/138 2026-03-09T19:49:57.987 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pyasn1-0.4.8-7.el9.noarch 17/138 2026-03-09T19:49:57.994 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-packaging-20.9-5.el9.noarch 18/138 2026-03-09T19:49:58.005 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 19/138 2026-03-09T19:49:58.013 INFO:teuthology.orchestra.run.vm11.stdout: Installing : protobuf-3.14.0-17.el9.x86_64 20/138 2026-03-09T19:49:58.017 INFO:teuthology.orchestra.run.vm11.stdout: Installing : lua-5.4.4-4.el9.x86_64 21/138 2026-03-09T19:49:58.022 INFO:teuthology.orchestra.run.vm11.stdout: Installing : flexiblas-3.0.4-9.el9.x86_64 22/138 2026-03-09T19:49:58.050 INFO:teuthology.orchestra.run.vm11.stdout: Installing : unzip-6.0-59.el9.x86_64 23/138 2026-03-09T19:49:58.066 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 24/138 2026-03-09T19:49:58.071 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-requests-2.25.1-10.el9.noarch 25/138 2026-03-09T19:49:58.078 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libquadmath-11.5.0-14.el9.x86_64 26/138 2026-03-09T19:49:58.082 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libgfortran-11.5.0-14.el9.x86_64 27/138 2026-03-09T19:49:58.113 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 28/138 2026-03-09T19:49:58.120 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 29/138 2026-03-09T19:49:58.131 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 30/138 2026-03-09T19:49:58.145 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 31/138 2026-03-09T19:49:58.154 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 32/138 2026-03-09T19:49:58.183 INFO:teuthology.orchestra.run.vm11.stdout: Installing : zip-3.0-35.el9.x86_64 33/138 2026-03-09T19:49:58.189 INFO:teuthology.orchestra.run.vm11.stdout: Installing : luarocks-3.9.2-5.el9.noarch 34/138 2026-03-09T19:49:58.198 INFO:teuthology.orchestra.run.vm11.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 35/138 2026-03-09T19:49:58.228 INFO:teuthology.orchestra.run.vm11.stdout: Installing : protobuf-compiler-3.14.0-17.el9.x86_64 36/138 2026-03-09T19:49:58.288 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-mako-1.1.4-6.el9.noarch 37/138 2026-03-09T19:49:58.307 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9.noarch 38/138 2026-03-09T19:49:58.315 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-rsa-4.9-2.el9.noarch 39/138 2026-03-09T19:49:58.325 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 40/138 2026-03-09T19:49:58.333 INFO:teuthology.orchestra.run.vm11.stdout: Installing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 41/138 2026-03-09T19:49:58.337 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 42/138 2026-03-09T19:49:58.355 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 43/138 2026-03-09T19:49:58.383 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 44/138 2026-03-09T19:49:58.392 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-webob-1.8.8-2.el9.noarch 45/138 2026-03-09T19:49:58.399 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 46/138 2026-03-09T19:49:58.414 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 47/138 2026-03-09T19:49:58.439 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 48/138 2026-03-09T19:49:58.451 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 49/138 2026-03-09T19:49:58.516 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-logutils-0.3.5-21.el9.noarch 50/138 2026-03-09T19:49:58.525 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pecan-1.4.2-3.el9.noarch 51/138 2026-03-09T19:49:58.535 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 52/138 2026-03-09T19:49:58.585 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 53/138 2026-03-09T19:49:58.961 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 54/138 2026-03-09T19:49:58.977 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 55/138 2026-03-09T19:49:58.983 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 56/138 2026-03-09T19:49:58.990 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 57/138 2026-03-09T19:49:58.995 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 58/138 2026-03-09T19:49:59.003 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 59/138 2026-03-09T19:49:59.007 INFO:teuthology.orchestra.run.vm11.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 60/138 2026-03-09T19:49:59.009 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 61/138 2026-03-09T19:49:59.040 INFO:teuthology.orchestra.run.vm11.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 62/138 2026-03-09T19:49:59.090 INFO:teuthology.orchestra.run.vm11.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 63/138 2026-03-09T19:49:59.105 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 64/138 2026-03-09T19:49:59.113 INFO:teuthology.orchestra.run.vm11.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 65/138 2026-03-09T19:49:59.119 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-toml-0.10.2-6.el9.noarch 66/138 2026-03-09T19:49:59.127 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 67/138 2026-03-09T19:49:59.133 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 68/138 2026-03-09T19:49:59.142 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 69/138 2026-03-09T19:49:59.148 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 70/138 2026-03-09T19:49:59.182 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 71/138 2026-03-09T19:49:59.194 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-protobuf-3.14.0-17.el9.noarch 72/138 2026-03-09T19:49:59.236 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 73/138 2026-03-09T19:49:59.511 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-devel-3.9.25-3.el9.x86_64 74/138 2026-03-09T19:49:59.543 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 75/138 2026-03-09T19:49:59.551 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jinja2-2.11.3-8.el9.noarch 76/138 2026-03-09T19:49:59.613 INFO:teuthology.orchestra.run.vm11.stdout: Installing : openblas-0.3.29-1.el9.x86_64 77/138 2026-03-09T19:49:59.616 INFO:teuthology.orchestra.run.vm11.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 78/138 2026-03-09T19:49:59.641 INFO:teuthology.orchestra.run.vm11.stdout: Installing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 79/138 2026-03-09T19:50:00.034 INFO:teuthology.orchestra.run.vm11.stdout: Installing : flexiblas-netlib-3.0.4-9.el9.x86_64 80/138 2026-03-09T19:50:00.135 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-numpy-1:1.23.5-2.el9.x86_64 81/138 2026-03-09T19:50:00.950 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 82/138 2026-03-09T19:50:00.978 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 83/138 2026-03-09T19:50:00.984 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libxslt-1.1.34-12.el9.x86_64 84/138 2026-03-09T19:50:00.989 INFO:teuthology.orchestra.run.vm11.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 85/138 2026-03-09T19:50:01.151 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 86/138 2026-03-09T19:50:01.153 INFO:teuthology.orchestra.run.vm11.stdout: Upgrading : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:50:01.187 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 87/138 2026-03-09T19:50:01.190 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 88/138 2026-03-09T19:50:01.198 INFO:teuthology.orchestra.run.vm11.stdout: Installing : boost-program-options-1.75.0-13.el9.x86_64 89/138 2026-03-09T19:50:01.470 INFO:teuthology.orchestra.run.vm11.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 90/138 2026-03-09T19:50:01.479 INFO:teuthology.orchestra.run.vm11.stdout: Installing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:50:01.507 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 91/138 2026-03-09T19:50:01.510 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 92/138 2026-03-09T19:50:03.005 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:50:03.009 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:50:03.033 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 93/138 2026-03-09T19:50:03.050 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-ply-3.11-14.el9.noarch 94/138 2026-03-09T19:50:03.071 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 95/138 2026-03-09T19:50:03.162 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 96/138 2026-03-09T19:50:03.176 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cryptography-36.0.1-5.el9.x86_64 97/138 2026-03-09T19:50:03.205 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 98/138 2026-03-09T19:50:03.242 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cheroot-10.0.1-4.el9.noarch 99/138 2026-03-09T19:50:03.305 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-cherrypy-18.6.1-2.el9.noarch 100/138 2026-03-09T19:50:03.316 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 101/138 2026-03-09T19:50:03.322 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:50:03.329 INFO:teuthology.orchestra.run.vm11.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 103/138 2026-03-09T19:50:03.334 INFO:teuthology.orchestra.run.vm11.stdout: Installing : qatlib-25.08.0-2.el9.x86_64 104/138 2026-03-09T19:50:03.335 INFO:teuthology.orchestra.run.vm11.stdout: Installing : qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:50:03.355 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 105/138 2026-03-09T19:50:03.685 INFO:teuthology.orchestra.run.vm11.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 106/138 2026-03-09T19:50:03.691 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:50:03.744 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 107/138 2026-03-09T19:50:03.744 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-03-09T19:50:03.744 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-03-09T19:50:03.744 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:03.750 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 108/138 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /sys 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /proc 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /mnt 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /var/tmp 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /home 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /root 2026-03-09T19:50:10.694 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /tmp 2026-03-09T19:50:10.695 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:10.836 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 109/138 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-03-09T19:50:10.866 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:11.100 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 110/138 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-03-09T19:50:11.125 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:11.133 INFO:teuthology.orchestra.run.vm11.stdout: Installing : mailcap-2.1.49-5.el9.noarch 111/138 2026-03-09T19:50:11.136 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 112/138 2026-03-09T19:50:11.154 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:50:11.154 INFO:teuthology.orchestra.run.vm11.stdout:Creating group 'qat' with GID 994. 2026-03-09T19:50:11.154 INFO:teuthology.orchestra.run.vm11.stdout:Creating group 'libstoragemgmt' with GID 993. 2026-03-09T19:50:11.154 INFO:teuthology.orchestra.run.vm11.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 993 and GID 993. 2026-03-09T19:50:11.154 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:11.167 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:50:11.195 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 113/138 2026-03-09T19:50:11.196 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-03-09T19:50:11.196 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:11.239 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 114/138 2026-03-09T19:50:11.318 INFO:teuthology.orchestra.run.vm11.stdout: Installing : cryptsetup-2.8.1-3.el9.x86_64 115/138 2026-03-09T19:50:11.323 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:50:11.339 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 116/138 2026-03-09T19:50:11.339 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:11.339 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:50:11.339 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:12.173 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 117/138 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-03-09T19:50:12.200 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:12.272 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:50:12.278 INFO:teuthology.orchestra.run.vm11.stdout: Installing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 118/138 2026-03-09T19:50:12.297 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 119/138 2026-03-09T19:50:12.321 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 120/138 2026-03-09T19:50:12.324 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:50:12.941 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 121/138 2026-03-09T19:50:12.948 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:50:13.545 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 122/138 2026-03-09T19:50:13.548 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:50:13.612 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 123/138 2026-03-09T19:50:13.675 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 124/138 2026-03-09T19:50:13.677 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 125/138 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-03-09T19:50:13.702 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:13.717 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:50:13.727 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 126/138 2026-03-09T19:50:14.265 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 127/138 2026-03-09T19:50:14.269 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:50:14.293 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 128/138 2026-03-09T19:50:14.294 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:14.294 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:50:14.294 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:50:14.294 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-03-09T19:50:14.294 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:14.305 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:50:14.338 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 129/138 2026-03-09T19:50:14.338 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:14.338 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:50:14.338 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:14.499 INFO:teuthology.orchestra.run.vm11.stdout: Installing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 130/138 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-03-09T19:50:14.525 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:17.211 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 131/138 2026-03-09T19:50:17.223 INFO:teuthology.orchestra.run.vm11.stdout: Installing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 132/138 2026-03-09T19:50:17.228 INFO:teuthology.orchestra.run.vm11.stdout: Installing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 133/138 2026-03-09T19:50:17.286 INFO:teuthology.orchestra.run.vm11.stdout: Installing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 134/138 2026-03-09T19:50:17.296 INFO:teuthology.orchestra.run.vm11.stdout: Installing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:50:17.300 INFO:teuthology.orchestra.run.vm11.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 136/138 2026-03-09T19:50:17.300 INFO:teuthology.orchestra.run.vm11.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:50:17.320 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 137/138 2026-03-09T19:50:17.320 INFO:teuthology.orchestra.run.vm11.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:50:18.701 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:50:18.701 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/138 2026-03-09T19:50:18.701 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/138 2026-03-09T19:50:18.701 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/138 2026-03-09T19:50:18.701 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 4/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 6/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 7/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 9/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 10/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 11/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 12/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_6 13/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 14/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 15/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 16/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 17/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 18/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9 19/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 20/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 21/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 22/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 23/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 24/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 25/138 2026-03-09T19:50:18.702 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 26/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 27/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 28/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 29/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 30/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 31/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 32/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 33/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 34/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 35/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 36/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 37/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 38/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 39/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 40/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 41/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 42/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 43/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 45/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ply-3.11-14.el9.noarch 46/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 47/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 48/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 49/138 2026-03-09T19:50:18.703 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : unzip-6.0-59.el9.x86_64 50/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : zip-3.0-35.el9.x86_64 51/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 52/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 53/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 54/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 55/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 56/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 57/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 58/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 59/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 60/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 61/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 62/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lua-5.4.4-4.el9.x86_64 63/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 64/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 65/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 66/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 67/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 68/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 69/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 70/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 71/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 72/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 73/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 74/138 2026-03-09T19:50:18.704 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 75/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 76/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 77/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 78/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 79/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 80/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 81/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 82/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 83/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 84/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 85/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 86/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 87/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 88/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 89/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 90/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 91/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 92/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 93/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 94/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 95/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 96/138 2026-03-09T19:50:18.705 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 97/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 98/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 99/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 100/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 101/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 102/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 103/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 104/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 105/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 106/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 107/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 108/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 109/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 110/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 111/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 112/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 113/138 2026-03-09T19:50:18.706 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 114/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 115/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 116/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 117/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 118/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 119/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 120/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 121/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 122/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 123/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 124/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 125/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 126/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 127/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 128/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 129/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 130/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 131/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 132/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : re2-1:20211101-20.el9.x86_64 133/138 2026-03-09T19:50:18.707 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 134/138 2026-03-09T19:50:18.708 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 135/138 2026-03-09T19:50:18.708 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 136/138 2026-03-09T19:50:18.708 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 137/138 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 138/138 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout:Upgraded: 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout:Installed: 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.816 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:50:18.817 INFO:teuthology.orchestra.run.vm11.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: mailcap-2.1.49-5.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-jmespath-1.0.1-1.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:50:18.818 INFO:teuthology.orchestra.run.vm11.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:50:18.819 INFO:teuthology.orchestra.run.vm11.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: socat-1.7.4.1-8.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:50:18.820 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:50:18.926 DEBUG:teuthology.parallel:result is None 2026-03-09T19:50:18.926 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:19.541 DEBUG:teuthology.orchestra.run.vm02:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T19:50:19.562 INFO:teuthology.orchestra.run.vm02.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:19.563 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:19.563 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T19:50:19.564 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:20.153 DEBUG:teuthology.orchestra.run.vm10:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T19:50:20.176 INFO:teuthology.orchestra.run.vm10.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:20.176 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:20.176 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T19:50:20.177 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:20.794 DEBUG:teuthology.orchestra.run.vm11:> rpm -q ceph --qf '%{VERSION}-%{RELEASE}' 2026-03-09T19:50:20.815 INFO:teuthology.orchestra.run.vm11.stdout:19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:20.815 INFO:teuthology.packaging:The installed version of ceph is 19.2.3-678.ge911bdeb.el9 2026-03-09T19:50:20.815 INFO:teuthology.task.install:The correct ceph version 19.2.3-678.ge911bdeb is installed. 2026-03-09T19:50:20.816 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-09T19:50:20.816 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:50:20.816 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T19:50:20.846 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:50:20.846 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T19:50:20.877 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:50:20.877 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-09T19:50:20.908 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-09T19:50:20.908 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:50:20.908 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T19:50:20.933 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T19:50:20.997 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:50:20.997 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T19:50:21.023 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T19:50:21.087 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:50:21.087 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/usr/bin/daemon-helper 2026-03-09T19:50:21.111 DEBUG:teuthology.orchestra.run.vm11:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-09T19:50:21.179 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-09T19:50:21.179 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:50:21.179 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T19:50:21.209 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T19:50:21.281 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:50:21.281 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T19:50:21.309 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T19:50:21.378 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:50:21.378 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-09T19:50:21.407 DEBUG:teuthology.orchestra.run.vm11:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-09T19:50:21.475 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-09T19:50:21.475 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:50:21.475 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T19:50:21.507 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T19:50:21.579 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:50:21.579 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T19:50:21.606 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T19:50:21.673 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:50:21.673 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/usr/bin/stdin-killer 2026-03-09T19:50:21.698 DEBUG:teuthology.orchestra.run.vm11:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-09T19:50:21.763 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon election default strategy': 1}, 'mgr': {'debug mgr': 20, 'debug ms': 1, 'mgr/cephadm/use_agent': False}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN', 'mons down', 'mon down', 'out of quorum', 'CEPHADM_STRAY_HOST', 'CEPHADM_STRAY_DAEMON', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Cluster fsid is 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.102', 'mon.b': '192.168.123.110', 'mon.c': '192.168.123.111'} 2026-03-09T19:50:21.806 INFO:tasks.cephadm:First mon is mon.a on vm02 2026-03-09T19:50:21.806 INFO:tasks.cephadm:First mgr is a 2026-03-09T19:50:21.806 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-09T19:50:21.807 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-09T19:50:21.838 DEBUG:teuthology.orchestra.run.vm10:> sudo hostname $(hostname -s) 2026-03-09T19:50:21.862 DEBUG:teuthology.orchestra.run.vm11:> sudo hostname $(hostname -s) 2026-03-09T19:50:21.886 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-09T19:50:21.886 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:22.499 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-09T19:50:23.077 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:50:23.078 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T19:50:23.078 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-09T19:50:23.078 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:24.632 INFO:teuthology.orchestra.run.vm02.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 19:50 /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:24.633 DEBUG:teuthology.orchestra.run.vm10:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:26.166 INFO:teuthology.orchestra.run.vm10.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 19:50 /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:26.166 DEBUG:teuthology.orchestra.run.vm11:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:27.766 INFO:teuthology.orchestra.run.vm11.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 9 19:50 /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:27.766 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:27.781 DEBUG:teuthology.orchestra.run.vm10:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:27.796 DEBUG:teuthology.orchestra.run.vm11:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-09T19:50:27.816 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-09T19:50:27.816 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T19:50:27.823 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T19:50:27.838 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-09T19:50:27.990 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T19:50:27.994 INFO:teuthology.orchestra.run.vm10.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T19:50:28.012 INFO:teuthology.orchestra.run.vm11.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout:{ 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout: "repo_digests": [ 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T19:51:27.145 INFO:teuthology.orchestra.run.vm10.stdout: ] 2026-03-09T19:51:27.146 INFO:teuthology.orchestra.run.vm10.stdout:} 2026-03-09T19:51:28.232 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-09T19:51:28.233 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout:{ 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout: "repo_digests": [ 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout: ] 2026-03-09T19:51:30.110 INFO:teuthology.orchestra.run.vm11.stdout:} 2026-03-09T19:51:30.131 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-09T19:51:30.160 DEBUG:teuthology.orchestra.run.vm10:> sudo mkdir -p /etc/ceph 2026-03-09T19:51:30.189 DEBUG:teuthology.orchestra.run.vm11:> sudo mkdir -p /etc/ceph 2026-03-09T19:51:30.220 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-09T19:51:30.244 DEBUG:teuthology.orchestra.run.vm10:> sudo chmod 777 /etc/ceph 2026-03-09T19:51:30.267 DEBUG:teuthology.orchestra.run.vm11:> sudo chmod 777 /etc/ceph 2026-03-09T19:51:30.291 INFO:tasks.cephadm:Writing seed config... 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [global] mon election default strategy = 1 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mgr] mgr/cephadm/use_agent = False 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-09T19:51:30.291 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-09T19:51:30.292 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:51:30.292 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-09T19:51:30.306 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd mon election default strategy = 1 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 mgr/cephadm/use_agent = False [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-09T19:51:30.306 DEBUG:teuthology.orchestra.run.vm02:mon.a> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a.service 2026-03-09T19:51:30.349 DEBUG:teuthology.orchestra.run.vm02:mgr.a> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a.service 2026-03-09T19:51:30.391 INFO:tasks.cephadm:Bootstrapping... 2026-03-09T19:51:30.392 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id a --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.102 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:51:30.542 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-09T19:51:30.542 INFO:teuthology.orchestra.run.vm02.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', '3527d82a-1bf1-11f1-92a4-57f58e2eb8dd', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'a', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.102', '--skip-admin-label'] 2026-03-09T19:51:30.543 INFO:teuthology.orchestra.run.vm02.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-09T19:51:30.543 INFO:teuthology.orchestra.run.vm02.stdout:Verifying podman|docker is present... 2026-03-09T19:51:30.569 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 5.8.0 2026-03-09T19:51:30.569 INFO:teuthology.orchestra.run.vm02.stdout:Verifying lvm2 is present... 2026-03-09T19:51:30.569 INFO:teuthology.orchestra.run.vm02.stdout:Verifying time synchronization is in place... 2026-03-09T19:51:30.577 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T19:51:30.585 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T19:51:30.585 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T19:51:30.585 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-09T19:51:30.590 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-09T19:51:30.596 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-09T19:51:30.596 INFO:teuthology.orchestra.run.vm02.stdout:Unit chronyd.service is enabled and running 2026-03-09T19:51:30.596 INFO:teuthology.orchestra.run.vm02.stdout:Repeating the final host check... 2026-03-09T19:51:30.619 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 5.8.0 2026-03-09T19:51:30.619 INFO:teuthology.orchestra.run.vm02.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-09T19:51:30.620 INFO:teuthology.orchestra.run.vm02.stdout:systemctl is present 2026-03-09T19:51:30.620 INFO:teuthology.orchestra.run.vm02.stdout:lvcreate is present 2026-03-09T19:51:30.626 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-09T19:51:30.626 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-09T19:51:30.633 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-09T19:51:30.633 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-09T19:51:30.641 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:Unit chronyd.service is enabled and running 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:Host looks OK 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:Cluster fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:Acquiring lock 139813296839984 on /run/cephadm/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.lock 2026-03-09T19:51:30.648 INFO:teuthology.orchestra.run.vm02.stdout:Lock 139813296839984 acquired on /run/cephadm/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.lock 2026-03-09T19:51:30.649 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 3300 ... 2026-03-09T19:51:30.649 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 6789 ... 2026-03-09T19:51:30.649 INFO:teuthology.orchestra.run.vm02.stdout:Base mon IP(s) is [192.168.123.102:3300, 192.168.123.102:6789], mon addrv is [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-09T19:51:30.654 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.102 metric 100 2026-03-09T19:51:30.654 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.102 metric 100 2026-03-09T19:51:30.657 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-09T19:51:30.657 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:2/64 scope link noprefixroute 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-09T19:51:30.660 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-09T19:51:30.661 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-09T19:51:30.661 INFO:teuthology.orchestra.run.vm02.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-09T19:51:30.661 INFO:teuthology.orchestra.run.vm02.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-09T19:51:30.662 INFO:teuthology.orchestra.run.vm02.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Getting image source signatures 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-09T19:51:32.279 INFO:teuthology.orchestra.run.vm02.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-09T19:51:32.470 INFO:teuthology.orchestra.run.vm02.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T19:51:32.470 INFO:teuthology.orchestra.run.vm02.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-09T19:51:32.470 INFO:teuthology.orchestra.run.vm02.stdout:Extracting ceph user uid/gid from container image... 2026-03-09T19:51:32.591 INFO:teuthology.orchestra.run.vm02.stdout:stat: stdout 167 167 2026-03-09T19:51:32.591 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial keys... 2026-03-09T19:51:32.699 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQDEJK9poUUtJxAAs1YTjdhcqpwyAsiqlzDnuQ== 2026-03-09T19:51:32.793 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQDEJK9pb9gOLhAATWOfd7Aj0upxNDqUt9X3PQ== 2026-03-09T19:51:32.884 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQDEJK9pTEK7MxAA13lzrurCPjCshsfsCu6EEQ== 2026-03-09T19:51:32.884 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial monmap... 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool for a [v2:192.168.123.102:3300,v1:192.168.123.102:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = quincy 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: set fsid to 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:32.994 INFO:teuthology.orchestra.run.vm02.stdout:Creating mon... 2026-03-09T19:51:33.149 INFO:teuthology.orchestra.run.vm02.stdout:create mon.a on 2026-03-09T19:51:33.300 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-03-09T19:51:33.420 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-09T19:51:33.551 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target → /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target. 2026-03-09T19:51:33.551 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target → /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target. 2026-03-09T19:51:33.702 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a 2026-03-09T19:51:33.702 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a.service: Unit ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a.service not loaded. 2026-03-09T19:51:33.844 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target.wants/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a.service → /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@.service. 2026-03-09T19:51:34.020 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-09T19:51:34.020 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T19:51:34.020 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon to start... 2026-03-09T19:51:34.020 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon... 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout cluster: 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout id: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout services: 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.157117s) 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout data: 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pgs: 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:mon is available 2026-03-09T19:51:34.285 INFO:teuthology.orchestra.run.vm02.stdout:Assimilating anything we can from ceph.conf... 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/cephadm/use_agent = False 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T19:51:34.495 INFO:teuthology.orchestra.run.vm02.stdout:Generating new minimal ceph.conf... 2026-03-09T19:51:34.687 INFO:teuthology.orchestra.run.vm02.stdout:Restarting the monitor... 2026-03-09T19:51:35.216 INFO:teuthology.orchestra.run.vm02.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-09T19:51:35.438 INFO:teuthology.orchestra.run.vm02.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-09T19:51:35.440 INFO:teuthology.orchestra.run.vm02.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:51:35.440 INFO:teuthology.orchestra.run.vm02.stdout:Creating mgr... 2026-03-09T19:51:35.440 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-09T19:51:35.441 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-09T19:51:35.582 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a 2026-03-09T19:51:35.582 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a.service: Unit ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a.service not loaded. 2026-03-09T19:51:35.726 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd.target.wants/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a.service → /etc/systemd/system/ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@.service. 2026-03-09T19:51:35.920 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:35 vm02 systemd[1]: Starting Ceph mgr.a for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr to start... 2026-03-09T19:51:36.166 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr... 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:35 vm02 podman[51833]: 2026-03-09 19:51:35.834247382 +0000 UTC m=+0.010189847 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:35 vm02 podman[51833]: 2026-03-09 19:51:35.968427911 +0000 UTC m=+0.144370365 container create 4afc1cfb894a0da947a1d01f2054434c6b6b2826152460fd4190c045201a1c7d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, CEPH_REF=squid, ceph=True) 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 podman[51833]: 2026-03-09 19:51:36.087840087 +0000 UTC m=+0.263782541 container init 4afc1cfb894a0da947a1d01f2054434c6b6b2826152460fd4190c045201a1c7d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 podman[51833]: 2026-03-09 19:51:36.090260071 +0000 UTC m=+0.266202525 container start 4afc1cfb894a0da947a1d01f2054434c6b6b2826152460fd4190c045201a1c7d (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.vendor=CentOS, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2) 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 bash[51833]: 4afc1cfb894a0da947a1d01f2054434c6b6b2826152460fd4190c045201a1c7d 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 systemd[1]: Started Ceph mgr.a for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:51:36.207 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:36.171+0000 7fcc1bcca140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T19:51:36.437 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "3527d82a-1bf1-11f1-92a4-57f58e2eb8dd", 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:36.438 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.439 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T19:51:34:062651+0000", 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T19:51:34.063350+0000", 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:36.440 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (1/15)... 2026-03-09T19:51:36.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:36.220+0000 7fcc1bcca140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T19:51:37.019 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:36.676+0000 7fcc1bcca140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T19:51:37.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.018+0000 7fcc1bcca140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: from numpy import show_config as show_numpy_config 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.110+0000 7fcc1bcca140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.149+0000 7fcc1bcca140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T19:51:37.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.223+0000 7fcc1bcca140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T19:51:38.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.745+0000 7fcc1bcca140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T19:51:38.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.851+0000 7fcc1bcca140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:38.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.890+0000 7fcc1bcca140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T19:51:38.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.923+0000 7fcc1bcca140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:38.001 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:37 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:37.964+0000 7fcc1bcca140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T19:51:38.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.000+0000 7fcc1bcca140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T19:51:38.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.176+0000 7fcc1bcca140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T19:51:38.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.229+0000 7fcc1bcca140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:38.751 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.485+0000 7fcc1bcca140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "3527d82a-1bf1-11f1-92a4-57f58e2eb8dd", 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T19:51:38.775 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T19:51:38.776 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T19:51:34:062651+0000", 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T19:51:34.063350+0000", 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:38.777 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (2/15)... 2026-03-09T19:51:39.024 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.821+0000 7fcc1bcca140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T19:51:39.025 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.859+0000 7fcc1bcca140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T19:51:39.025 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.902+0000 7fcc1bcca140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T19:51:39.025 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:38 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:38.985+0000 7fcc1bcca140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T19:51:39.025 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:39.023+0000 7fcc1bcca140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T19:51:39.369 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:39.103+0000 7fcc1bcca140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T19:51:39.369 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:39.221+0000 7fcc1bcca140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:39.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:39.368+0000 7fcc1bcca140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T19:51:39.670 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:39.410+0000 7fcc1bcca140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T19:51:40.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:39 vm02 ceph-mon[51633]: from='mgr.14100 192.168.123.102:0/1410649966' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "3527d82a-1bf1-11f1-92a4-57f58e2eb8dd", 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-09T19:51:41.055 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-09T19:51:34:062651+0000", 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T19:51:41.056 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "restful" 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-09T19:51:34.063350+0000", 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:41.057 INFO:teuthology.orchestra.run.vm02.stdout:mgr is available 2026-03-09T19:51:41.310 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:41.310 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-09T19:51:41.310 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:41.310 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-09T19:51:41.310 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-09T19:51:41.311 INFO:teuthology.orchestra.run.vm02.stdout:Enabling cephadm module... 2026-03-09T19:51:41.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:41 vm02 ceph-mon[51633]: mgrmap e3: a(active, since 1.0109s) 2026-03-09T19:51:41.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:41 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2467587212' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-09T19:51:41.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:41 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1901414960' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-09T19:51:41.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:41 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1901414960' entity='client.admin' cmd='[{"prefix": "config assimilate-conf"}]': finished 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1315501315' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1315501315' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-mon[51633]: mgrmap e4: a(active, since 2s) 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: ignoring --setuser ceph since I am not root 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: ignoring --setgroup ceph since I am not root 2026-03-09T19:51:42.580 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:42.387+0000 7f487f48e140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T19:51:42.581 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:42.433+0000 7f487f48e140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 4, 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-09T19:51:42.620 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 4... 2026-03-09T19:51:43.166 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:42.863+0000 7f487f48e140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.166+0000 7f487f48e140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: from numpy import show_config as show_numpy_config 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.255+0000 7f487f48e140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T19:51:43.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.292+0000 7f487f48e140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T19:51:43.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.363+0000 7f487f48e140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T19:51:43.872 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/578674066' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T19:51:44.154 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.872+0000 7f487f48e140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T19:51:44.154 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:43 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:43.982+0000 7f487f48e140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:44.154 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.029+0000 7f487f48e140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T19:51:44.154 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.068+0000 7f487f48e140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:44.154 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.113+0000 7f487f48e140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T19:51:44.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.153+0000 7f487f48e140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T19:51:44.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.339+0000 7f487f48e140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T19:51:44.421 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.395+0000 7f487f48e140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:44.829 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.640+0000 7f487f48e140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T19:51:45.081 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.927+0000 7f487f48e140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T19:51:45.082 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:44.964+0000 7f487f48e140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T19:51:45.082 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.006+0000 7f487f48e140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T19:51:45.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.081+0000 7f487f48e140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T19:51:45.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.117+0000 7f487f48e140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T19:51:45.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.198+0000 7f487f48e140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T19:51:45.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.315+0000 7f487f48e140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:45.920 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.454+0000 7f487f48e140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T19:51:45.920 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:45.495+0000 7f487f48e140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: Active manager daemon a restarted 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: Activating manager daemon a 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: osdmap e2: 0 total, 0 up, 0 in 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: mgrmap e5: a(active, starting, since 0.00634091s) 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: Manager daemon a is now available 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:45 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 6, 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 4 is available 2026-03-09T19:51:46.543 INFO:teuthology.orchestra.run.vm02.stdout:Setting orchestrator backend to cephadm... 2026-03-09T19:51:46.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:46 vm02 ceph-mon[51633]: Found migration_current of "None". Setting to last migration. 2026-03-09T19:51:46.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:46 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T19:51:46.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:46 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:46.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:46 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:46.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:46 vm02 ceph-mon[51633]: mgrmap e6: a(active, since 1.00876s) 2026-03-09T19:51:47.089 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-09T19:51:47.089 INFO:teuthology.orchestra.run.vm02.stdout:Generating ssh key... 2026-03-09T19:51:47.600 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:46] ENGINE Bus STARTING 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:46] ENGINE Serving on https://192.168.123.102:7150 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:46] ENGINE Client ('192.168.123.102', 42310) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:46] ENGINE Serving on http://192.168.123.102:8765 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:46] ENGINE Bus STARTED 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='client.14130 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='client.14132 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: Generating ssh key... 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Generating public/private ed25519 key pair. 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Your identification has been saved in /tmp/tmph6pylxn_/key 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Your public key has been saved in /tmp/tmph6pylxn_/key.pub 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: The key fingerprint is: 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: SHA256:+h9qIfSs/Qw386l9DkhYgvgntQv9dWA53swktLzn5oo ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: The key's randomart image is: 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: +--[ED25519 256]--+ 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | . | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | . . o o | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | . . o . O . | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | ..o = o X | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | .+oS . + * | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | .=++ o + | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | .+o.B . o | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | ..o= B =. | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: | .ooE.=+o | 2026-03-09T19:51:47.601 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:47 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: +----[SHA256]-----+ 2026-03-09T19:51:47.648 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:47.648 INFO:teuthology.orchestra.run.vm02.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-09T19:51:47.648 INFO:teuthology.orchestra.run.vm02.stdout:Adding key to root@localhost authorized_keys... 2026-03-09T19:51:47.649 INFO:teuthology.orchestra.run.vm02.stdout:Adding host vm02... 2026-03-09T19:51:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:48 vm02 ceph-mon[51633]: from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:48 vm02 ceph-mon[51633]: from='client.14138 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:48 vm02 ceph-mon[51633]: Deploying cephadm binary to vm02 2026-03-09T19:51:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:48 vm02 ceph-mon[51633]: mgrmap e7: a(active, since 2s) 2026-03-09T19:51:49.414 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Added host 'vm02' with addr '192.168.123.102' 2026-03-09T19:51:49.414 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mon service... 2026-03-09T19:51:49.714 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-09T19:51:49.714 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mgr service... 2026-03-09T19:51:49.981 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: Added host vm02 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: Saving service mon spec with placement count:5 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: Saving service mgr spec with placement count:2 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:50.475 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:50 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/818076492' entity='client.admin' 2026-03-09T19:51:50.482 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the dashboard module... 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1188743793' entity='client.admin' 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/102650159' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-mon[51633]: from='mgr.14118 192.168.123.102:0/3662778175' entity='mgr.a' 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: ignoring --setuser ceph since I am not root 2026-03-09T19:51:51.617 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: ignoring --setgroup ceph since I am not root 2026-03-09T19:51:51.894 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:51.650+0000 7f9cb0c22140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T19:51:51.894 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:51 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:51.696+0000 7f9cb0c22140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T19:51:51.905 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 8, 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "a", 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-09T19:51:51.906 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 8... 2026-03-09T19:51:52.170 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:52.122+0000 7f9cb0c22140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T19:51:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/102650159' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-09T19:51:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-mon[51633]: mgrmap e8: a(active, since 5s) 2026-03-09T19:51:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/83897313' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:52.454+0000 7f9cb0c22140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: from numpy import show_config as show_numpy_config 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:52.548+0000 7f9cb0c22140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:52.587+0000 7f9cb0c22140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T19:51:52.921 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:52 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:52.660+0000 7f9cb0c22140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T19:51:53.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.150+0000 7f9cb0c22140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T19:51:53.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.260+0000 7f9cb0c22140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:53.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.299+0000 7f9cb0c22140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T19:51:53.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.338+0000 7f9cb0c22140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:53.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.380+0000 7f9cb0c22140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T19:51:53.898 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.423+0000 7f9cb0c22140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T19:51:53.898 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.611+0000 7f9cb0c22140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T19:51:53.898 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.668+0000 7f9cb0c22140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T19:51:54.170 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:53 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:53.897+0000 7f9cb0c22140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T19:51:54.459 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.178+0000 7f9cb0c22140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T19:51:54.459 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.217+0000 7f9cb0c22140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T19:51:54.459 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.258+0000 7f9cb0c22140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T19:51:54.459 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.335+0000 7f9cb0c22140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T19:51:54.459 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.374+0000 7f9cb0c22140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T19:51:54.744 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.458+0000 7f9cb0c22140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T19:51:54.744 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.571+0000 7f9cb0c22140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:51:54.744 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.707+0000 7f9cb0c22140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: Active manager daemon a restarted 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: Activating manager daemon a 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: osdmap e3: 0 total, 0 up, 0 in 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: mgrmap e9: a(active, starting, since 0.00673973s) 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "a", "id": "a"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: Manager daemon a is now available 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/mirror_snapshot_schedule"}]: dispatch 2026-03-09T19:51:55.171 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:51:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:51:54.743+0000 7f9cb0c22140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T19:51:55.819 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-09T19:51:55.819 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 10, 2026-03-09T19:51:55.819 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-09T19:51:55.819 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-09T19:51:55.819 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 8 is available 2026-03-09T19:51:55.820 INFO:teuthology.orchestra.run.vm02.stdout:Generating a dashboard self-signed certificate... 2026-03-09T19:51:56.174 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/a/trash_purge_schedule"}]: dispatch 2026-03-09T19:51:56.174 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:51:56.174 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:55 vm02 ceph-mon[51633]: mgrmap e10: a(active, since 1.01274s) 2026-03-09T19:51:56.320 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-09T19:51:56.320 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial admin user... 2026-03-09T19:51:56.734 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$vaOcTqzwh8Kq3P9ghgG4luMTIIAvk0ETYjLkcr87dvLK496sqrj6e", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773085916, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-09T19:51:56.734 INFO:teuthology.orchestra.run.vm02.stdout:Fetching dashboard port number... 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:56] ENGINE Bus STARTING 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='client.14162 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:56] ENGINE Serving on https://192.168.123.102:7150 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:56] ENGINE Client ('192.168.123.102', 36742) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:56] ENGINE Serving on http://192.168.123.102:8765 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: [09/Mar/2026:19:51:56] ENGINE Bus STARTED 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='client.14164 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:51:57.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:51:57.640 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 8443 2026-03-09T19:51:57.640 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-09T19:51:57.640 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout:Ceph Dashboard is now available at: 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout: URL: https://vm02.local:8443/ 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout: User: admin 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout: Password: 0aqca17q4r 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:57.642 INFO:teuthology.orchestra.run.vm02.stdout:Saving cluster configuration to /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config directory 2026-03-09T19:51:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:57 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1563140251' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-09T19:51:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:57 vm02 ceph-mon[51633]: mgrmap e11: a(active, since 2s) 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:Or, if you are only running a single cluster on this host: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: ceph telemetry on 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:For more information see: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:51:58.045 INFO:teuthology.orchestra.run.vm02.stdout:Bootstrap complete. 2026-03-09T19:51:58.077 INFO:tasks.cephadm:Fetching config... 2026-03-09T19:51:58.077 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:51:58.077 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-09T19:51:58.096 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-09T19:51:58.096 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:51:58.096 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-09T19:51:58.157 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-09T19:51:58.157 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:51:58.157 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/keyring of=/dev/stdout 2026-03-09T19:51:58.224 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-09T19:51:58.224 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:51:58.224 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-09T19:51:58.282 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-09T19:51:58.282 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T19:51:58.376 INFO:teuthology.orchestra.run.vm02.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:58.397 DEBUG:teuthology.orchestra.run.vm10:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T19:51:58.433 INFO:teuthology.orchestra.run.vm10.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:58.445 DEBUG:teuthology.orchestra.run.vm11:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-09T19:51:58.483 INFO:teuthology.orchestra.run.vm11.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIA842XzxKLgcA0zXfyswV9VPQCZabsNVrT0z4WhQmDRO ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:51:58.496 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-09T19:51:58.697 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:51:59.025 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-09T19:51:59.026 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-09T19:51:59.229 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:51:59.254 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:58 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2783840950' entity='client.admin' 2026-03-09T19:51:59.254 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:51:58 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2875165813' entity='client.admin' 2026-03-09T19:51:59.514 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm10 2026-03-09T19:51:59.514 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:51:59.514 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.conf 2026-03-09T19:51:59.531 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:51:59.531 DEBUG:teuthology.orchestra.run.vm10:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:51:59.587 INFO:tasks.cephadm:Adding host vm10 to orchestrator... 2026-03-09T19:51:59.587 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch host add vm10 2026-03-09T19:51:59.756 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:00.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:00.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:00.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='client.14172 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:00.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:00 vm02 ceph-mon[51633]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: Updating vm02:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm10", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: Updating vm02:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.client.admin.keyring 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:01.639 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:01 vm02 ceph-mon[51633]: Deploying cephadm binary to vm10 2026-03-09T19:52:01.640 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm10' with addr '192.168.123.110' 2026-03-09T19:52:01.694 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch host ls --format=json 2026-03-09T19:52:01.876 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:02.110 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:52:02.110 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.110", "hostname": "vm10", "labels": [], "status": ""}] 2026-03-09T19:52:02.180 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm11 2026-03-09T19:52:02.180 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:52:02.180 DEBUG:teuthology.orchestra.run.vm11:> dd of=/etc/ceph/ceph.conf 2026-03-09T19:52:02.197 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:52:02.197 DEBUG:teuthology.orchestra.run.vm11:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:02.255 INFO:tasks.cephadm:Adding host vm11 to orchestrator... 2026-03-09T19:52:02.255 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch host add vm11 2026-03-09T19:52:02.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:02 vm02 ceph-mon[51633]: mgrmap e12: a(active, since 6s) 2026-03-09T19:52:02.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:02 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:02.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:02 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:02.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:02 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:02.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:02 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:02.426 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:03.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:03 vm02 ceph-mon[51633]: Added host vm10 2026-03-09T19:52:03.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:03 vm02 ceph-mon[51633]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:52:03.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:04.264 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm11' with addr '192.168.123.111' 2026-03-09T19:52:04.312 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch host ls --format=json 2026-03-09T19:52:04.496 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm11", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: Deploying cephadm binary to vm11 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm10", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:04.522 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:04.737 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:52:04.738 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.110", "hostname": "vm10", "labels": [], "status": ""}, {"addr": "192.168.123.111", "hostname": "vm11", "labels": [], "status": ""}] 2026-03-09T19:52:04.825 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-09T19:52:04.825 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd crush tunables default 2026-03-09T19:52:05.007 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:05.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: Added host vm11 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: Updating vm10:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: Updating vm10:/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:05.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:05 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/4046926994' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-09T19:52:06.192 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-09T19:52:06.241 INFO:tasks.cephadm:Adding mon.a on vm02 2026-03-09T19:52:06.241 INFO:tasks.cephadm:Adding mon.b on vm10 2026-03-09T19:52:06.241 INFO:tasks.cephadm:Adding mon.c on vm11 2026-03-09T19:52:06.241 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch apply mon '3;vm02:192.168.123.102=a;vm10:192.168.123.110=b;vm11:192.168.123.111=c' 2026-03-09T19:52:06.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:06 vm02 ceph-mon[51633]: Updating vm10:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.client.admin.keyring 2026-03-09T19:52:06.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:06 vm02 ceph-mon[51633]: from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:52:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:06 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/4046926994' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-09T19:52:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:06 vm02 ceph-mon[51633]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:06.425 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T19:52:06.478 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T19:52:06.735 INFO:teuthology.orchestra.run.vm11.stdout:Scheduled mon update... 2026-03-09T19:52:06.792 DEBUG:teuthology.orchestra.run.vm10:mon.b> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.b.service 2026-03-09T19:52:06.793 DEBUG:teuthology.orchestra.run.vm11:mon.c> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.c.service 2026-03-09T19:52:06.795 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T19:52:06.795 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph mon dump -f json 2026-03-09T19:52:07.071 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T19:52:07.113 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-09T19:52:07.387 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:52:07.387 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","modified":"2026-03-09T19:51:32.974864Z","created":"2026-03-09T19:51:32.974864Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T19:52:07.389 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T19:52:08.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm10:192.168.123.110=b;vm11:192.168.123.111=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Saving service mon spec with placement vm02:192.168.123.102=a;vm10:192.168.123.110=b;vm11:192.168.123.111=c;count:3 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd/host:vm11", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Updating vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Updating vm11:/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/3779300897' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Updating vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.client.admin.keyring 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:08.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:07 vm02 ceph-mon[51633]: Deploying daemon mon.c on vm11 2026-03-09T19:52:08.463 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T19:52:08.463 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph mon dump -f json 2026-03-09T19:52:08.735 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:52:09.041 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:52:09.041 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":1,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","modified":"2026-03-09T19:51:32.974864Z","created":"2026-03-09T19:51:32.974864Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-09T19:52:09.041 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 1 2026-03-09T19:52:10.093 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-09T19:52:10.094 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph mon dump -f json 2026-03-09T19:52:10.269 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:52:10.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:10 vm10 ceph-mon[55027]: mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-09T19:52:14.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:14.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: Deploying daemon mon.b on vm10 2026-03-09T19:52:14.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: mon.a calling monitor election 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: mon.c calling monitor election 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: monmap epoch 2 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: last_changed 2026-03-09T19:52:09.046814+0000 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: min_mon_release 19 (squid) 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: election_strategy: 1 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: fsmap 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: mgrmap e12: a(active, since 19s) 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: overall HEALTH_OK 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:15.420 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:52:15 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:52:15.045+0000 7f9c7cf81640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-09T19:52:19.742 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:52:19.742 INFO:teuthology.orchestra.run.vm11.stdout:{"epoch":3,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","modified":"2026-03-09T19:52:14.453821Z","created":"2026-03-09T19:51:32.974864Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:3300","nonce":0},{"type":"v1","addr":"192.168.123.111:6789","nonce":0}]},"addr":"192.168.123.111:6789/0","public_addr":"192.168.123.111:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:3300","nonce":0},{"type":"v1","addr":"192.168.123.110:6789","nonce":0}]},"addr":"192.168.123.110:6789/0","public_addr":"192.168.123.110:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-09T19:52:19.742 INFO:teuthology.orchestra.run.vm11.stderr:dumped monmap epoch 3 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: mon.a calling monitor election 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: mon.c calling monitor election 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: monmap epoch 3 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: min_mon_release 19 (squid) 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: election_strategy: 1 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: fsmap 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: mgrmap e12: a(active, since 24s) 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: overall HEALTH_OK 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:19.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:19 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:19.815 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-09T19:52:19.815 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph config generate-minimal-conf 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: mon.a calling monitor election 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: mon.c calling monitor election 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: monmap epoch 3 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: min_mon_release 19 (squid) 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: election_strategy: 1 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: fsmap 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: mgrmap e12: a(active, since 24s) 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: overall HEALTH_OK 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.845 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.846 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.846 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.846 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:19.846 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:19.846 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:19 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:20.019 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:20.303 INFO:teuthology.orchestra.run.vm02.stdout:# minimal ceph.conf for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:20.304 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-09T19:52:20.304 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:20.304 INFO:teuthology.orchestra.run.vm02.stdout: mon_host = [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] 2026-03-09T19:52:20.355 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-09T19:52:20.355 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:52:20.355 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T19:52:20.423 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:52:20.423 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:20.498 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:52:20.498 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T19:52:20.525 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:52:20.525 DEBUG:teuthology.orchestra.run.vm10:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:20.589 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:52:20.589 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/ceph/ceph.conf 2026-03-09T19:52:20.624 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:52:20.624 DEBUG:teuthology.orchestra.run.vm11:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:52:20.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/913719532' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/103307789' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.698 INFO:tasks.cephadm:Adding mgr.a on vm02 2026-03-09T19:52:20.698 INFO:tasks.cephadm:Adding mgr.b on vm10 2026-03-09T19:52:20.698 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch apply mgr '2;vm02=a;vm10=b' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='client.? 192.168.123.111:0/913719532' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/103307789' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:20.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:20.904 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:52:21.152 INFO:teuthology.orchestra.run.vm11.stdout:Scheduled mgr update... 2026-03-09T19:52:21.224 DEBUG:teuthology.orchestra.run.vm10:mgr.b> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.b.service 2026-03-09T19:52:21.225 INFO:tasks.cephadm:Deploying OSDs... 2026-03-09T19:52:21.225 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-09T19:52:21.225 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T19:52:21.241 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:52:21.241 DEBUG:teuthology.orchestra.run.vm02:> ls /dev/[sv]d? 2026-03-09T19:52:21.296 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vda 2026-03-09T19:52:21.297 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdb 2026-03-09T19:52:21.297 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdc 2026-03-09T19:52:21.297 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vdd 2026-03-09T19:52:21.297 INFO:teuthology.orchestra.run.vm02.stdout:/dev/vde 2026-03-09T19:52:21.297 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T19:52:21.297 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T19:52:21.297 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdb 2026-03-09T19:52:21.356 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdb 2026-03-09T19:52:21.356 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:21.356 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-09 19:51:58.965250660 +0000 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-09 19:48:18.723335100 +0000 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-09 19:48:18.723335100 +0000 2026-03-09T19:52:21.357 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-09 19:35:27.247000000 +0000 2026-03-09T19:52:21.357 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T19:52:21.425 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-09T19:52:21.425 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-09T19:52:21.425 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000180457 s, 2.8 MB/s 2026-03-09T19:52:21.426 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T19:52:21.492 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdc 2026-03-09T19:52:21.512 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: Deploying daemon mon.b on vm10 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.a calling monitor election 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.c calling monitor election 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: monmap epoch 2 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: last_changed 2026-03-09T19:52:09.046814+0000 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: min_mon_release 19 (squid) 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: election_strategy: 1 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: fsmap 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mgrmap e12: a(active, since 19s) 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: overall HEALTH_OK 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.a calling monitor election 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.c calling monitor election 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:21.513 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: monmap epoch 3 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: min_mon_release 19 (squid) 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: election_strategy: 1 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: fsmap 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: mgrmap e12: a(active, since 24s) 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: overall HEALTH_OK 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='client.? 192.168.123.111:0/913719532' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/103307789' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:21.514 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:21.515 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.515 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:21.515 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:52:21.515 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:52:21.515 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdc 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-09 19:51:59.014250679 +0000 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-09 19:48:18.666335038 +0000 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-09 19:48:18.666335038 +0000 2026-03-09T19:52:21.549 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-09 19:35:27.268000000 +0000 2026-03-09T19:52:21.550 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T19:52:21.613 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-09T19:52:21.613 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-09T19:52:21.613 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000211185 s, 2.4 MB/s 2026-03-09T19:52:21.614 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T19:52:21.670 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vdd 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vdd 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-09 19:51:59.052250694 +0000 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-09 19:48:18.713335090 +0000 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-09 19:48:18.713335090 +0000 2026-03-09T19:52:21.728 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-09 19:35:27.279000000 +0000 2026-03-09T19:52:21.729 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T19:52:21.792 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-09T19:52:21.792 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-09T19:52:21.792 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000230671 s, 2.2 MB/s 2026-03-09T19:52:21.794 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T19:52:21.803 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 systemd[1]: Starting Ceph mgr.b for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:52:21.850 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vde 2026-03-09T19:52:21.907 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vde 2026-03-09T19:52:21.907 INFO:teuthology.orchestra.run.vm02.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-09 19:51:59.084250706 +0000 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-09 19:48:18.686335060 +0000 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-09 19:48:18.686335060 +0000 2026-03-09T19:52:21.908 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-03-09 19:35:27.282000000 +0000 2026-03-09T19:52:21.908 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T19:52:21.976 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-09T19:52:21.976 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-09T19:52:21.977 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000223117 s, 2.3 MB/s 2026-03-09T19:52:21.978 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T19:52:22.037 DEBUG:teuthology.orchestra.run.vm10:> set -ex 2026-03-09T19:52:22.037 DEBUG:teuthology.orchestra.run.vm10:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T19:52:22.059 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:52:22.059 DEBUG:teuthology.orchestra.run.vm10:> ls /dev/[sv]d? 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 podman[55950]: 2026-03-09 19:52:21.917340902 +0000 UTC m=+0.019727600 container create 314b77a7d3a1d88b6cb2f1b7c8eb0effa2a09aedf8321724bbfc200cb8dab0f0 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 podman[55950]: 2026-03-09 19:52:21.960573507 +0000 UTC m=+0.062960205 container init 314b77a7d3a1d88b6cb2f1b7c8eb0effa2a09aedf8321724bbfc200cb8dab0f0 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default) 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 podman[55950]: 2026-03-09 19:52:21.96382837 +0000 UTC m=+0.066215068 container start 314b77a7d3a1d88b6cb2f1b7c8eb0effa2a09aedf8321724bbfc200cb8dab0f0 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b, OSD_FLAVOR=default, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True) 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 bash[55950]: 314b77a7d3a1d88b6cb2f1b7c8eb0effa2a09aedf8321724bbfc200cb8dab0f0 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 podman[55950]: 2026-03-09 19:52:21.909097652 +0000 UTC m=+0.011484350 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:21 vm10 systemd[1]: Started Ceph mgr.b for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:52:22.086 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:22.080+0000 7f42665be140 -1 mgr[py] Module alerts has missing NOTIFY_TYPES member 2026-03-09T19:52:22.113 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vda 2026-03-09T19:52:22.113 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdb 2026-03-09T19:52:22.113 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdc 2026-03-09T19:52:22.113 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vdd 2026-03-09T19:52:22.113 INFO:teuthology.orchestra.run.vm10.stdout:/dev/vde 2026-03-09T19:52:22.113 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T19:52:22.113 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T19:52:22.113 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdb 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdb 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 19:52:04.065945233 +0000 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 19:48:18.800305646 +0000 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 19:48:18.800305646 +0000 2026-03-09T19:52:22.235 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 19:34:56.253000000 +0000 2026-03-09T19:52:22.235 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T19:52:22.282 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T19:52:22.282 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T19:52:22.282 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000197449 s, 2.6 MB/s 2026-03-09T19:52:22.284 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T19:52:22.386 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdc 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: mon.b calling monitor election 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: Deploying daemon mgr.b on vm10 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: mon.b calling monitor election 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: mon.a calling monitor election 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: monmap epoch 3 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: min_mon_release 19 (squid) 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: election_strategy: 1 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: fsmap 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: mgrmap e12: a(active, since 26s) 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: overall HEALTH_OK 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:22 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdc 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 19:52:04.102945276 +0000 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 19:48:18.786305631 +0000 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 19:48:18.786305631 +0000 2026-03-09T19:52:22.423 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 19:34:56.262000000 +0000 2026-03-09T19:52:22.424 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T19:52:22.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: mon.b calling monitor election 2026-03-09T19:52:22.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: Deploying daemon mgr.b on vm10 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: mon.b calling monitor election 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: mon.a calling monitor election 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: monmap epoch 3 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: min_mon_release 19 (squid) 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: election_strategy: 1 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: fsmap 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: mgrmap e12: a(active, since 26s) 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: overall HEALTH_OK 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:22.463 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:22.125+0000 7f42665be140 -1 mgr[py] Module balancer has missing NOTIFY_TYPES member 2026-03-09T19:52:22.530 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T19:52:22.530 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T19:52:22.530 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000185488 s, 2.8 MB/s 2026-03-09T19:52:22.532 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T19:52:22.569 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vdd 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: mon.b calling monitor election 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='client.14205 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "2;vm02=a;vm10=b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: Saving service mgr spec with placement vm02=a;vm10=b;count:2 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: Deploying daemon mgr.b on vm10 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: mon.b calling monitor election 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: mon.a calling monitor election 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: monmap epoch 3 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: last_changed 2026-03-09T19:52:14.453821+0000 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: min_mon_release 19 (squid) 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: election_strategy: 1 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: 1: [v2:192.168.123.111:3300/0,v1:192.168.123.111:6789/0] mon.c 2026-03-09T19:52:22.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: 2: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: fsmap 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: osdmap e4: 0 total, 0 up, 0 in 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: mgrmap e12: a(active, since 26s) 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: overall HEALTH_OK 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:22.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:22 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:22.685 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vdd 2026-03-09T19:52:22.685 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:22.685 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-09T19:52:22.685 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:22.686 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:22.686 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 19:52:04.133945312 +0000 2026-03-09T19:52:22.686 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 19:48:18.815305663 +0000 2026-03-09T19:52:22.686 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 19:48:18.815305663 +0000 2026-03-09T19:52:22.686 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 19:34:56.273000000 +0000 2026-03-09T19:52:22.686 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T19:52:22.719 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T19:52:22.719 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T19:52:22.719 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000185326 s, 2.8 MB/s 2026-03-09T19:52:22.721 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T19:52:22.784 DEBUG:teuthology.orchestra.run.vm10:> stat /dev/vde 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout: File: /dev/vde 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Access: 2026-03-09 19:52:04.157945340 +0000 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Modify: 2026-03-09 19:48:18.833305683 +0000 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout:Change: 2026-03-09 19:48:18.833305683 +0000 2026-03-09T19:52:22.843 INFO:teuthology.orchestra.run.vm10.stdout: Birth: 2026-03-09 19:34:56.280000000 +0000 2026-03-09T19:52:22.843 DEBUG:teuthology.orchestra.run.vm10:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T19:52:22.912 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:22.624+0000 7f42665be140 -1 mgr[py] Module crash has missing NOTIFY_TYPES member 2026-03-09T19:52:22.915 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records in 2026-03-09T19:52:22.915 INFO:teuthology.orchestra.run.vm10.stderr:1+0 records out 2026-03-09T19:52:22.915 INFO:teuthology.orchestra.run.vm10.stderr:512 bytes copied, 0.000324939 s, 1.6 MB/s 2026-03-09T19:52:22.917 DEBUG:teuthology.orchestra.run.vm10:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T19:52:22.978 DEBUG:teuthology.orchestra.run.vm11:> set -ex 2026-03-09T19:52:22.978 DEBUG:teuthology.orchestra.run.vm11:> dd if=/scratch_devs of=/dev/stdout 2026-03-09T19:52:22.995 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:52:22.995 DEBUG:teuthology.orchestra.run.vm11:> ls /dev/[sv]d? 2026-03-09T19:52:23.050 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vda 2026-03-09T19:52:23.050 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdb 2026-03-09T19:52:23.050 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdc 2026-03-09T19:52:23.050 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vdd 2026-03-09T19:52:23.050 INFO:teuthology.orchestra.run.vm11.stdout:/dev/vde 2026-03-09T19:52:23.050 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-09T19:52:23.050 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-09T19:52:23.050 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdb 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdb 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 225 Links: 1 Device type: fc,10 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 19:52:06.720021357 +0000 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 19:48:19.181322817 +0000 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 19:48:19.181322817 +0000 2026-03-09T19:52:23.107 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 19:34:20.279000000 +0000 2026-03-09T19:52:23.107 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:22 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:22.997+0000 7f42665be140 -1 mgr[py] Module devicehealth has missing NOTIFY_TYPES member 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-09T19:52:23.166 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: from numpy import show_config as show_numpy_config 2026-03-09T19:52:23.167 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.089+0000 7f42665be140 -1 mgr[py] Module diskprediction_local has missing NOTIFY_TYPES member 2026-03-09T19:52:23.167 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.125+0000 7f42665be140 -1 mgr[py] Module influx has missing NOTIFY_TYPES member 2026-03-09T19:52:23.174 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T19:52:23.174 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T19:52:23.174 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000127468 s, 4.0 MB/s 2026-03-09T19:52:23.176 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-09T19:52:23.234 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdc 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdc 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 228 Links: 1 Device type: fc,20 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 19:52:06.775021414 +0000 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 19:48:19.178322814 +0000 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 19:48:19.178322814 +0000 2026-03-09T19:52:23.293 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 19:34:20.283000000 +0000 2026-03-09T19:52:23.293 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-09T19:52:23.357 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T19:52:23.357 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T19:52:23.357 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000171982 s, 3.0 MB/s 2026-03-09T19:52:23.359 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-09T19:52:23.416 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vdd 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:52:23.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.463 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.192+0000 7f42665be140 -1 mgr[py] Module iostat has missing NOTIFY_TYPES member 2026-03-09T19:52:23.474 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vdd 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 230 Links: 1 Device type: fc,30 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 19:52:06.815021456 +0000 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 19:48:19.201322840 +0000 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 19:48:19.201322840 +0000 2026-03-09T19:52:23.475 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 19:34:20.288000000 +0000 2026-03-09T19:52:23.475 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:23.538 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:23.540 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T19:52:23.540 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T19:52:23.540 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000146884 s, 3.5 MB/s 2026-03-09T19:52:23.541 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-09T19:52:23.599 DEBUG:teuthology.orchestra.run.vm11:> stat /dev/vde 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout: File: /dev/vde 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Device: 6h/6d Inode: 233 Links: 1 Device type: fc,40 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Access: 2026-03-09 19:52:06.853021496 +0000 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Modify: 2026-03-09 19:48:19.258322907 +0000 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout:Change: 2026-03-09 19:48:19.258322907 +0000 2026-03-09T19:52:23.658 INFO:teuthology.orchestra.run.vm11.stdout: Birth: 2026-03-09 19:34:20.292000000 +0000 2026-03-09T19:52:23.658 DEBUG:teuthology.orchestra.run.vm11:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-09T19:52:23.723 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records in 2026-03-09T19:52:23.723 INFO:teuthology.orchestra.run.vm11.stderr:1+0 records out 2026-03-09T19:52:23.723 INFO:teuthology.orchestra.run.vm11.stderr:512 bytes copied, 0.000170759 s, 3.0 MB/s 2026-03-09T19:52:23.724 DEBUG:teuthology.orchestra.run.vm11:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-09T19:52:23.782 INFO:tasks.cephadm:Deploying osd.0 on vm02 with /dev/vde... 2026-03-09T19:52:23.782 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vde 2026-03-09T19:52:23.803 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:52:23 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:52:23.453+0000 7f9c7cf81640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-03-09T19:52:23.943 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:23.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.690+0000 7f42665be140 -1 mgr[py] Module nfs has missing NOTIFY_TYPES member 2026-03-09T19:52:23.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.795+0000 7f42665be140 -1 mgr[py] Module orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:52:23.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.836+0000 7f42665be140 -1 mgr[py] Module osd_perf_query has missing NOTIFY_TYPES member 2026-03-09T19:52:23.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.868+0000 7f42665be140 -1 mgr[py] Module osd_support has missing NOTIFY_TYPES member 2026-03-09T19:52:23.945 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.907+0000 7f42665be140 -1 mgr[py] Module pg_autoscaler has missing NOTIFY_TYPES member 2026-03-09T19:52:24.212 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-mon[55027]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T19:52:24.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-mon[55027]: Reconfiguring daemon mgr.a on vm02 2026-03-09T19:52:24.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-mon[55027]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:24.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:23 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:23.944+0000 7f42665be140 -1 mgr[py] Module progress has missing NOTIFY_TYPES member 2026-03-09T19:52:24.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.120+0000 7f42665be140 -1 mgr[py] Module prometheus has missing NOTIFY_TYPES member 2026-03-09T19:52:24.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.171+0000 7f42665be140 -1 mgr[py] Module rbd_support has missing NOTIFY_TYPES member 2026-03-09T19:52:24.344 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:24 vm02 ceph-mon[51633]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T19:52:24.344 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:24 vm02 ceph-mon[51633]: Reconfiguring daemon mgr.a on vm02 2026-03-09T19:52:24.344 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:24 vm02 ceph-mon[51633]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:24.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:24 vm11 ceph-mon[76771]: Reconfiguring mgr.a (unknown last config time)... 2026-03-09T19:52:24.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:24 vm11 ceph-mon[76771]: Reconfiguring daemon mgr.a on vm02 2026-03-09T19:52:24.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:24 vm11 ceph-mon[76771]: pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:24.679 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.402+0000 7f42665be140 -1 mgr[py] Module rgw has missing NOTIFY_TYPES member 2026-03-09T19:52:24.728 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:52:24.754 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm02:/dev/vde 2026-03-09T19:52:24.943 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.678+0000 7f42665be140 -1 mgr[py] Module rook has missing NOTIFY_TYPES member 2026-03-09T19:52:24.943 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.715+0000 7f42665be140 -1 mgr[py] Module selftest has missing NOTIFY_TYPES member 2026-03-09T19:52:24.943 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.758+0000 7f42665be140 -1 mgr[py] Module snap_schedule has missing NOTIFY_TYPES member 2026-03-09T19:52:24.943 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.832+0000 7f42665be140 -1 mgr[py] Module status has missing NOTIFY_TYPES member 2026-03-09T19:52:24.943 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.866+0000 7f42665be140 -1 mgr[py] Module telegraf has missing NOTIFY_TYPES member 2026-03-09T19:52:24.966 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:25.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:24 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:24.941+0000 7f42665be140 -1 mgr[py] Module telemetry has missing NOTIFY_TYPES member 2026-03-09T19:52:25.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:25.052+0000 7f42665be140 -1 mgr[py] Module test_orchestrator has missing NOTIFY_TYPES member 2026-03-09T19:52:25.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:25.185+0000 7f42665be140 -1 mgr[py] Module volumes has missing NOTIFY_TYPES member 2026-03-09T19:52:25.486 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-b[55960]: 2026-03-09T19:52:25.221+0000 7f42665be140 -1 mgr[py] Module zabbix has missing NOTIFY_TYPES member 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: Standby manager daemon b started 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T19:52:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:25 vm02 ceph-mon[51633]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: Standby manager daemon b started 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T19:52:25.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:25 vm11 ceph-mon[76771]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T19:52:25.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: Standby manager daemon b started 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/crt"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/b/key"}]: dispatch 2026-03-09T19:52:25.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:25 vm10 ceph-mon[55027]: from='mgr.? 192.168.123.110:0/1875581709' entity='mgr.b' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: mgrmap e13: a(active, since 30s), standbys: b 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]: dispatch 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]': finished 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:26.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:26 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2559750758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: mgrmap e13: a(active, since 30s), standbys: b 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]: dispatch 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]': finished 2026-03-09T19:52:26.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T19:52:26.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:26.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:26 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2559750758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='client.14217 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: mgrmap e13: a(active, since 30s), standbys: b 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr metadata", "who": "b", "id": "b"}]: dispatch 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]: dispatch 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/434558942' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5bf6a389-7f23-4b97-8b5b-9b8f72542309"}]': finished 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: osdmap e5: 1 total, 0 up, 1 in 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:26.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:26 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2559750758' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:28.765 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:28 vm10 ceph-mon[55027]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:28.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:28 vm11 ceph-mon[76771]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:28.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:28 vm02 ceph-mon[51633]: pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:30.609 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:30 vm02 ceph-mon[51633]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:30.609 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:30 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T19:52:30.609 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:30 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:30.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:30 vm11 ceph-mon[76771]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:30.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:30 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T19:52:30.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:30 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:30.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:30 vm10 ceph-mon[55027]: pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:30.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:30 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T19:52:30.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:30 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:31.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:31 vm11 ceph-mon[76771]: Deploying daemon osd.0 on vm02 2026-03-09T19:52:31.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:31 vm02 ceph-mon[51633]: Deploying daemon osd.0 on vm02 2026-03-09T19:52:31.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:31 vm10 ceph-mon[55027]: Deploying daemon osd.0 on vm02 2026-03-09T19:52:32.776 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:32 vm11 ceph-mon[76771]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:32.777 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:32 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:32.777 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:32 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:32.777 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:32 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:32.788 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:32 vm02 ceph-mon[51633]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:32.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:32 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:32.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:32 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:32.789 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:32 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:32.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:32 vm10 ceph-mon[55027]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:32.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:32 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:32.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:32 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:32.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:32 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:33.558 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 0 on host 'vm02' 2026-03-09T19:52:33.613 DEBUG:teuthology.orchestra.run.vm02:osd.0> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.0.service 2026-03-09T19:52:33.616 INFO:tasks.cephadm:Deploying osd.1 on vm02 with /dev/vdd... 2026-03-09T19:52:33.616 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vdd 2026-03-09T19:52:33.901 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:33 vm11 ceph-mon[76771]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:33 vm02 ceph-mon[51633]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:34.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:33 vm10 ceph-mon[55027]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-09T19:52:35.208 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:52:35.231 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm02:/dev/vdd 2026-03-09T19:52:35.408 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: Detected new or changed devices on vm02 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm02 to 257.0M 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: Unable to set osd_memory_target on vm02 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T19:52:35.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T19:52:35.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:35.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:35 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: Detected new or changed devices on vm02 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm02 to 257.0M 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: Unable to set osd_memory_target on vm02 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:35.894 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:35 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:35.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: Detected new or changed devices on vm02 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm02 to 257.0M 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: Unable to set osd_memory_target on vm02 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: osdmap e6: 1 total, 0 up, 1 in 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:35.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:35 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.561 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:36.562 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:52:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0[61227]: 2026-03-09T19:52:36.456+0000 7fd809807640 -1 osd.0 0 waiting for initial osdmap 2026-03-09T19:52:36.562 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:52:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0[61227]: 2026-03-09T19:52:36.461+0000 7fd80461d640 -1 osd.0 7 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]': finished 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959] boot 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:36 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]': finished 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959] boot 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:36 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: osdmap e7: 1 total, 0 up, 1 in 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959]' entity='osd.0' 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2272695357' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "34047d7e-0449-4e8b-bb14-2f79ca4f35d2"}]': finished 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: osd.0 [v2:192.168.123.102:6802/1639950959,v1:192.168.123.102:6803/1639950959] boot 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: osdmap e8: 2 total, 1 up, 2 in 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-09T19:52:36.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:36 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:37 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:52:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:37 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:52:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:37 vm11 ceph-mon[76771]: from='client.14241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:37 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3938635952' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:37.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:37 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:52:37.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:37 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:52:37.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:37 vm02 ceph-mon[51633]: from='client.14241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:37.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:37 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3938635952' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:37.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:37 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:52:37.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:37 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:52:37.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:37 vm10 ceph-mon[55027]: from='client.14241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:37.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:37 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3938635952' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:38.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:38 vm11 ceph-mon[76771]: pgmap v19: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:38.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:38 vm11 ceph-mon[76771]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T19:52:38.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:38 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:38.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:38 vm02 ceph-mon[51633]: pgmap v19: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:38.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:38 vm02 ceph-mon[51633]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T19:52:38.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:38 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:38.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:38 vm10 ceph-mon[55027]: pgmap v19: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:38.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:38 vm10 ceph-mon[55027]: osdmap e9: 2 total, 1 up, 2 in 2026-03-09T19:52:38.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:38 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:40.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:40 vm11 ceph-mon[76771]: pgmap v21: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:40.895 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:40 vm02 ceph-mon[51633]: pgmap v21: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:40.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:40 vm10 ceph-mon[55027]: pgmap v21: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:41.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T19:52:41.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:41.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:41 vm02 ceph-mon[51633]: Deploying daemon osd.1 on vm02 2026-03-09T19:52:41.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:41 vm02 ceph-mon[51633]: pgmap v22: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:41.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:41 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T19:52:41.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:41 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:41.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:41 vm11 ceph-mon[76771]: Deploying daemon osd.1 on vm02 2026-03-09T19:52:41.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:41 vm11 ceph-mon[76771]: pgmap v22: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T19:52:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:41 vm10 ceph-mon[55027]: Deploying daemon osd.1 on vm02 2026-03-09T19:52:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:41 vm10 ceph-mon[55027]: pgmap v22: 0 pgs: ; 0 B data, 50 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.827 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:43.915 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:43.970 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 1 on host 'vm02' 2026-03-09T19:52:44.031 DEBUG:teuthology.orchestra.run.vm02:osd.1> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.1.service 2026-03-09T19:52:44.033 INFO:tasks.cephadm:Deploying osd.2 on vm10 with /dev/vde... 2026-03-09T19:52:44.033 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vde 2026-03-09T19:52:44.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:44.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:43 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:44.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.214 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.b/config 2026-03-09T19:52:44.829 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:52:44 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:52:44.529+0000 7fdd0179e740 -1 osd.1 0 log_to_monitors true 2026-03-09T19:52:44.830 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:44 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:44.830 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:44 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.830 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:44 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.830 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:44 vm02 ceph-mon[51633]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T19:52:44.904 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:44 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:44.905 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:44 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.905 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:44 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:44.905 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:44 vm10 ceph-mon[55027]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T19:52:45.010 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:52:45.027 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm10:/dev/vde 2026-03-09T19:52:45.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:44 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:45.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:44 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:45.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:44 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:45.124 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:44 vm11 ceph-mon[76771]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-09T19:52:45.199 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.b/config 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: Detected new or changed devices on vm02 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm02 to 128.5M 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: Unable to set osd_memory_target on vm02 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:46.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:46.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:46.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: Detected new or changed devices on vm02 2026-03-09T19:52:46.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm02 to 128.5M 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: Unable to set osd_memory_target on vm02 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:46.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:45 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: Detected new or changed devices on vm02 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm02 to 128.5M 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: Unable to set osd_memory_target on vm02 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: osdmap e10: 2 total, 1 up, 2 in 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='client.24149 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:46.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='client.? 192.168.123.110:0/2321246754' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]': finished 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: osdmap e12: 3 total, 1 up, 3 in 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='client.? 192.168.123.110:0/2876277443' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:47.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:46 vm02 ceph-mon[51633]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' 2026-03-09T19:52:47.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:52:46 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:52:46.899+0000 7fdcfd71f640 -1 osd.1 0 waiting for initial osdmap 2026-03-09T19:52:47.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:52:46 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:52:46.906+0000 7fdcf8d48640 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='client.? 192.168.123.110:0/2321246754' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]': finished 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: osdmap e12: 3 total, 1 up, 3 in 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='client.? 192.168.123.110:0/2876277443' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:47.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:46 vm11 ceph-mon[76771]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: osdmap e11: 2 total, 1 up, 2 in 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='client.? 192.168.123.110:0/2321246754' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "bd50b4cb-b42e-4544-8af8-f525f5e040d1"}]': finished 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: osdmap e12: 3 total, 1 up, 3 in 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='client.? 192.168.123.110:0/2876277443' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:47.440 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:46 vm10 ceph-mon[55027]: from='osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092]' entity='osd.1' 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092] boot 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:47 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092] boot 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: osd.1 [v2:192.168.123.102:6810/3659897092,v1:192.168.123.102:6811/3659897092] boot 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: osdmap e13: 3 total, 2 up, 3 in 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-09T19:52:48.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:50.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:49 vm10 ceph-mon[55027]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:50.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:49 vm10 ceph-mon[55027]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T19:52:50.213 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:50.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:49 vm11 ceph-mon[76771]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:50.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:49 vm11 ceph-mon[76771]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T19:52:50.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:49 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:50.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:49 vm02 ceph-mon[51633]: pgmap v30: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:50.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:49 vm02 ceph-mon[51633]: osdmap e14: 3 total, 2 up, 3 in 2026-03-09T19:52:50.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:51.034 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T19:52:51.034 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:51.034 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:50 vm10 ceph-mon[55027]: Deploying daemon osd.2 on vm10 2026-03-09T19:52:51.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:50 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T19:52:51.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:50 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:51.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:50 vm11 ceph-mon[76771]: Deploying daemon osd.2 on vm10 2026-03-09T19:52:51.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T19:52:51.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:51.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:50 vm02 ceph-mon[51633]: Deploying daemon osd.2 on vm10 2026-03-09T19:52:52.189 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:51 vm10 ceph-mon[55027]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:52.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:51 vm11 ceph-mon[76771]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:52.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:51 vm02 ceph-mon[51633]: pgmap v32: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:53.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:53 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.389 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:53.389 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.389 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.390 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.390 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.390 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:53.390 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:53.390 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:53.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:53.489 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 2 on host 'vm10' 2026-03-09T19:52:53.540 DEBUG:teuthology.orchestra.run.vm10:osd.2> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.2.service 2026-03-09T19:52:53.541 INFO:tasks.cephadm:Deploying osd.3 on vm10 with /dev/vdd... 2026-03-09T19:52:53.542 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vdd 2026-03-09T19:52:53.823 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.b/config 2026-03-09T19:52:54.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:54 vm02 ceph-mon[51633]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:54.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:54.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:54.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:54 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:54.713 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:52:54 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:52:54.538+0000 7f15bc309740 -1 osd.2 0 log_to_monitors true 2026-03-09T19:52:54.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:54 vm10 ceph-mon[55027]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:54.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:54 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:54.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:54 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:54.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:54 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:54.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:54 vm11 ceph-mon[76771]: pgmap v33: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:54.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:54 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:52:54.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:54 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:54.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:54 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.207 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:52:55.224 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm10:/dev/vdd 2026-03-09T19:52:55.397 INFO:teuthology.orchestra.run.vm10.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.b/config 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: Detected new or changed devices on vm10 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm10 to 257.0M 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: Unable to set osd_memory_target on vm10 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:55 vm10 ceph-mon[55027]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: Detected new or changed devices on vm10 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm10 to 257.0M 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: Unable to set osd_memory_target on vm10 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:55.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:55 vm11 ceph-mon[76771]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:55.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: Detected new or changed devices on vm10 2026-03-09T19:52:55.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm10 to 257.0M 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: Unable to set osd_memory_target on vm10 to 269530726: error parsing value: Value '269530726' is below minimum 939524096 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:55.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:55 vm02 ceph-mon[51633]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='client.? 192.168.123.110:0/2861215481' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]': finished 2026-03-09T19:52:56.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: osdmap e16: 4 total, 2 up, 4 in 2026-03-09T19:52:56.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:56.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:56 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='client.? 192.168.123.110:0/2861215481' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]': finished 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: osdmap e16: 4 total, 2 up, 4 in 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:56.887 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: osdmap e15: 3 total, 2 up, 3 in 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='client.? 192.168.123.110:0/2861215481' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "133b4719-ab0d-4d74-a537-a82bc3bfb22f"}]': finished 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: osdmap e16: 4 total, 2 up, 4 in 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:56.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:57.463 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:52:57 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:52:57.297+0000 7f15b828a640 -1 osd.2 0 waiting for initial osdmap 2026-03-09T19:52:57.463 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:52:57 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:52:57.305+0000 7f15b38b3640 -1 osd.2 16 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:52:57.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:57 vm11 ceph-mon[76771]: from='client.14283 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:57.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:57 vm11 ceph-mon[76771]: from='client.? 192.168.123.110:0/1532861610' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:57.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:57 vm11 ceph-mon[76771]: from='osd.2 ' entity='osd.2' 2026-03-09T19:52:57.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:57 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:57.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:57 vm02 ceph-mon[51633]: from='client.14283 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:57 vm02 ceph-mon[51633]: from='client.? 192.168.123.110:0/1532861610' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:57 vm02 ceph-mon[51633]: from='osd.2 ' entity='osd.2' 2026-03-09T19:52:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:57.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:57 vm10 ceph-mon[55027]: from='client.14283 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm10:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:52:57.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:57 vm10 ceph-mon[55027]: from='client.? 192.168.123.110:0/1532861610' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:52:57.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:57 vm10 ceph-mon[55027]: from='osd.2 ' entity='osd.2' 2026-03-09T19:52:57.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340] boot 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:58.822 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:58.864 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340] boot 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: pgmap v37: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: osd.2 [v2:192.168.123.110:6800/1532484340,v1:192.168.123.110:6801/1532484340] boot 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: osdmap e17: 4 total, 3 up, 4 in 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-09T19:52:58.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:58 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:52:59.624 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:52:59 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:52:59.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:52:59 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:52:59.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:52:59 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:00.523 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: pgmap v39: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T19:53:00.524 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: pgmap v39: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T19:53:00.850 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:00.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: pgmap v39: 0 pgs: ; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:00.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:00.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: osdmap e18: 4 total, 3 up, 4 in 2026-03-09T19:53:00.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:00.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:00.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T19:53:00.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:01.153 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63523]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T19:53:01.153 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63523]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.153 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63523]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.153 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 sudo[63523]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.154 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63501]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T19:53:01.154 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63501]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.154 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63501]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.154 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:53:00 vm10 sudo[63501]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69219]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T19:53:01.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69219]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69219]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69219]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69215]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vdd 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69215]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69215]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69215]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69211]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vde 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69211]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69211]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.171 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:53:00 vm02 sudo[69211]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 sudo[78218]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-09T19:53:01.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 sudo[78218]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-09T19:53:01.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 sudo[78218]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-09T19:53:01.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:00 vm11 sudo[78218]: pam_unix(sudo:session): session closed for user root 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: Deploying daemon osd.3 on vm10 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.714 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:01 vm10 ceph-mon[55027]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: Deploying daemon osd.3 on vm10 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:01 vm11 ceph-mon[76771]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: Deploying daemon osd.3 on vm10 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: osdmap e19: 4 total, 3 up, 4 in 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "c"}]: dispatch 2026-03-09T19:53:01.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:01 vm02 ceph-mon[51633]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-09T19:53:02.579 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:02 vm10 ceph-mon[55027]: pgmap v42: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:02.579 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:02 vm10 ceph-mon[55027]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T19:53:02.579 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:02 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:02.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:02 vm11 ceph-mon[76771]: pgmap v42: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:02.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:02 vm11 ceph-mon[76771]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T19:53:02.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:02 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:02.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:02 vm02 ceph-mon[51633]: pgmap v42: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:02.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:02 vm02 ceph-mon[51633]: osdmap e20: 4 total, 3 up, 4 in 2026-03-09T19:53:02.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:02 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:03.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:03 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:03.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:03 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:03 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:03 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:03 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:03.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:03 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.958 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:03 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:03.958 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:03 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.958 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:03 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.958 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:03 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:03.958 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:03 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.024 INFO:teuthology.orchestra.run.vm10.stdout:Created osd(s) 3 on host 'vm10' 2026-03-09T19:53:04.087 DEBUG:teuthology.orchestra.run.vm10:osd.3> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.3.service 2026-03-09T19:53:04.089 INFO:tasks.cephadm:Deploying osd.4 on vm11 with /dev/vde... 2026-03-09T19:53:04.089 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vde 2026-03-09T19:53:04.263 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:53:04.585 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:53:04 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:53:04.406+0000 7f8a370b6740 -1 osd.3 0 log_to_monitors true 2026-03-09T19:53:04.585 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: mgrmap e14: a(active, since 68s), standbys: b 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:04.586 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:04 vm10 ceph-mon[55027]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: mgrmap e14: a(active, since 68s), standbys: b 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:04.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:04.625 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:04 vm11 ceph-mon[76771]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: pgmap v44: 1 pgs: 1 unknown; 0 B data, 79 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: mgrmap e14: a(active, since 68s), standbys: b 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:04 vm02 ceph-mon[51633]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-09T19:53:05.110 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:53:05.134 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm11:/dev/vde 2026-03-09T19:53:05.308 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:53:06.035 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:06.035 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: Detected new or changed devices on vm10 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm10 to 128.5M 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: Unable to set osd_memory_target on vm10 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='client.24187 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:06.374 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:06 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: Detected new or changed devices on vm10 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm10 to 128.5M 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: Unable to set osd_memory_target on vm10 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='client.24187 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:06.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:06 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: pgmap v45: 1 pgs: 1 active+clean; 449 KiB data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: osdmap e21: 4 total, 3 up, 4 in 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: Detected new or changed devices on vm10 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm10 to 128.5M 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: Unable to set osd_memory_target on vm10 to 134765363: error parsing value: Value '134765363' is below minimum 939524096 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='client.24187 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vde", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:06 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:06.463 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:53:06 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:53:06.039+0000 7f8a33037640 -1 osd.3 0 waiting for initial osdmap 2026-03-09T19:53:06.463 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:53:06 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:53:06.046+0000 7f8a2ee61640 -1 osd.3 22 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='client.? 192.168.123.111:0/2968181173' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]': finished 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477] boot 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: osdmap e23: 5 total, 4 up, 5 in 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:07.198 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:07 vm11 ceph-mon[76771]: from='client.? 192.168.123.111:0/434750591' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:07.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/2968181173' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]': finished 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477] boot 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: osdmap e23: 5 total, 4 up, 5 in 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:07.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:07 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/434750591' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm10", "root=default"]}]': finished 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: osdmap e22: 4 total, 3 up, 4 in 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='client.? 192.168.123.111:0/2968181173' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "4bb90fd1-5e97-470b-b51e-8fefb71fa68d"}]': finished 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: osd.3 [v2:192.168.123.110:6808/3577442477,v1:192.168.123.110:6809/3577442477] boot 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: osdmap e23: 5 total, 4 up, 5 in 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:07.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:07 vm10 ceph-mon[55027]: from='client.? 192.168.123.111:0/434750591' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:08.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:08 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:53:08.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:08 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:08.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:08 vm02 ceph-mon[51633]: pgmap v49: 1 pgs: 1 remapped; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:08.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:08 vm02 ceph-mon[51633]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T19:53:08.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:08 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:08.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:08 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:53:08.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:08 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:08.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:08 vm10 ceph-mon[55027]: pgmap v49: 1 pgs: 1 remapped; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:08.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:08 vm10 ceph-mon[55027]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T19:53:08.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:08 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:08.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:08 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:53:08.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:08 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:53:08.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:08 vm11 ceph-mon[76771]: pgmap v49: 1 pgs: 1 remapped; 449 KiB data, 81 MiB used, 60 GiB / 60 GiB avail 2026-03-09T19:53:08.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:08 vm11 ceph-mon[76771]: osdmap e24: 5 total, 4 up, 5 in 2026-03-09T19:53:08.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:08 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:09.537 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:09 vm02 ceph-mon[51633]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T19:53:09.538 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:09 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:09.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:09 vm10 ceph-mon[55027]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T19:53:09.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:09 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:09.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:09 vm11 ceph-mon[76771]: osdmap e25: 5 total, 4 up, 5 in 2026-03-09T19:53:09.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:09 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:10.448 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:10 vm11 ceph-mon[76771]: pgmap v52: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:10.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:10 vm10 ceph-mon[55027]: pgmap v52: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:10.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:10 vm02 ceph-mon[51633]: pgmap v52: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:11.619 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:11 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T19:53:11.619 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:11 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:11.619 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:11 vm11 ceph-mon[76771]: Deploying daemon osd.4 on vm11 2026-03-09T19:53:11.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:11 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T19:53:11.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:11 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:11.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:11 vm02 ceph-mon[51633]: Deploying daemon osd.4 on vm11 2026-03-09T19:53:11.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:11 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-09T19:53:11.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:11 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:11.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:11 vm10 ceph-mon[55027]: Deploying daemon osd.4 on vm11 2026-03-09T19:53:12.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:12 vm02 ceph-mon[51633]: pgmap v53: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:12.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:12 vm10 ceph-mon[55027]: pgmap v53: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:12.773 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:12 vm11 ceph-mon[76771]: pgmap v53: 1 pgs: 1 remapped; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:13.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:13 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:13.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:13 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:13.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:13 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:13.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:13 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:13.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:13 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:13.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:13 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:13.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:13 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:13.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:13 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:13.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:13 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.156 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 4 on host 'vm11' 2026-03-09T19:53:14.235 DEBUG:teuthology.orchestra.run.vm11:osd.4> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service 2026-03-09T19:53:14.237 INFO:tasks.cephadm:Deploying osd.5 on vm11 with /dev/vdd... 2026-03-09T19:53:14.237 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- lvm zap /dev/vdd 2026-03-09T19:53:14.527 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.550 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:14 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:14 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 70 KiB/s, 0 objects/s recovering 2026-03-09T19:53:14.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:14.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:14 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.374 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:15 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:15.120+0000 7f33915a0740 -1 osd.4 0 log_to_monitors true 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:15.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:15 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:15.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:15 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:15.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:15 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:16.009 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:53:16.025 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph orch daemon add osd vm11:/dev/vdd 2026-03-09T19:53:16.197 INFO:teuthology.orchestra.run.vm11.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.c/config 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: Detected new or changed devices on vm11 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm11 to 4353M 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:16.821 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:16 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:16.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:16.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: Detected new or changed devices on vm11 2026-03-09T19:53:16.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm11 to 4353M 2026-03-09T19:53:16.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T19:53:16.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:16.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:16 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: Detected new or changed devices on vm11 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm11 to 4353M 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: osdmap e26: 5 total, 4 up, 5 in 2026-03-09T19:53:16.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:16.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='osd.4 ' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:16.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-09T19:53:16.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-09T19:53:16.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:16 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='client.? 192.168.123.111:0/3350835504' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]': finished 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: osdmap e28: 6 total, 4 up, 6 in 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:17.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:17 vm11 ceph-mon[76771]: from='osd.4 ' entity='osd.4' 2026-03-09T19:53:17.625 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:17 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:17.504+0000 7f338dd34640 -1 osd.4 0 waiting for initial osdmap 2026-03-09T19:53:17.625 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:17 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:17.511+0000 7f3388b4a640 -1 osd.4 28 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:53:17.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:17.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/3350835504' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]': finished 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: osdmap e28: 6 total, 4 up, 6 in 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:17.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:17 vm02 ceph-mon[51633]: from='osd.4 ' entity='osd.4' 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='client.24214 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm11:/dev/vdd", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: osdmap e27: 5 total, 4 up, 5 in 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='client.? 192.168.123.111:0/3350835504' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]: dispatch 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2b1993e3-b0a4-4651-a5df-7776ee00c4e5"}]': finished 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: osdmap e28: 6 total, 4 up, 6 in 2026-03-09T19:53:17.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:17.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:17.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:17 vm10 ceph-mon[55027]: from='osd.4 ' entity='osd.4' 2026-03-09T19:53:18.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: from='client.? 192.168.123.111:0/645025661' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984] boot 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:18 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: from='client.? 192.168.123.111:0/645025661' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984] boot 2026-03-09T19:53:18.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T19:53:18.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:18 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:18.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:53:18.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:18.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 107 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: from='client.? 192.168.123.111:0/645025661' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: osd.4 [v2:192.168.123.111:6800/3542558984,v1:192.168.123.111:6801/3542558984] boot 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: osdmap e29: 6 total, 5 up, 6 in 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-09T19:53:18.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:18 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:20.483 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:20 vm11 ceph-mon[76771]: pgmap v61: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:20.483 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:20 vm11 ceph-mon[76771]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T19:53:20.483 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:20 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:20 vm02 ceph-mon[51633]: pgmap v61: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:20 vm02 ceph-mon[51633]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T19:53:20.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:20 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:20.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:20 vm10 ceph-mon[55027]: pgmap v61: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:20.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:20 vm10 ceph-mon[55027]: osdmap e30: 6 total, 5 up, 6 in 2026-03-09T19:53:20.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:20 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:21.312 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:21 vm11 ceph-mon[76771]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T19:53:21.312 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:21 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:21.312 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:21 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:21.312 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:21 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:21.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:21 vm02 ceph-mon[51633]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T19:53:21.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:21 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:21.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:21 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:21.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:21 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:21.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:21 vm10 ceph-mon[55027]: osdmap e31: 6 total, 5 up, 6 in 2026-03-09T19:53:21.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:21.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:21.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:21 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:22.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:22 vm11 ceph-mon[76771]: pgmap v64: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:22.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:22 vm11 ceph-mon[76771]: Deploying daemon osd.5 on vm11 2026-03-09T19:53:22.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:22 vm02 ceph-mon[51633]: pgmap v64: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:22.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:22 vm02 ceph-mon[51633]: Deploying daemon osd.5 on vm11 2026-03-09T19:53:22.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:22 vm10 ceph-mon[55027]: pgmap v64: 1 pgs: 1 peering; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail 2026-03-09T19:53:22.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:22 vm10 ceph-mon[55027]: Deploying daemon osd.5 on vm11 2026-03-09T19:53:23.320 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:23.320 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:23.320 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:23 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:23.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:23.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:23.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:23 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:23.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:23.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:23.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:23 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.171 INFO:teuthology.orchestra.run.vm11.stdout:Created osd(s) 5 on host 'vm11' 2026-03-09T19:53:24.235 DEBUG:teuthology.orchestra.run.vm11:osd.5> sudo journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5.service 2026-03-09T19:53:24.236 INFO:tasks.cephadm:Waiting for 6 OSDs to come up... 2026-03-09T19:53:24.237 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd stat -f json 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail; 82 KiB/s, 0 objects/s recovering 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.375 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:24 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.414 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:24.533 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail; 82 KiB/s, 0 objects/s recovering 2026-03-09T19:53:24.533 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.534 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:24 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 533 MiB used, 99 GiB / 100 GiB avail; 82 KiB/s, 0 objects/s recovering 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:24 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:24.740 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":31,"num_osds":6,"num_up_osds":5,"osd_up_since":1773085998,"num_in_osds":6,"osd_in_since":1773085997,"num_remapped_pgs":0} 2026-03-09T19:53:25.333 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:25 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:25.052+0000 7f3d40d2e740 -1 osd.5 0 log_to_monitors true 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3746824085' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:25.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:25 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3746824085' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:25.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:25.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:25 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3746824085' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:25.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:25.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:25 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:25.741 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd stat -f json 2026-03-09T19:53:25.916 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:26.139 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:26.208 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":32,"num_osds":6,"num_up_osds":5,"osd_up_since":1773085998,"num_in_osds":6,"osd_in_since":1773085997,"num_remapped_pgs":0} 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: Detected new or changed devices on vm11 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: Adjusting osd_memory_target on vm11 to 2176M 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:26 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2256813375' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: Detected new or changed devices on vm11 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: Adjusting osd_memory_target on vm11 to 2176M 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:26 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2256813375' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:26.624 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:26 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:26.345+0000 7f3d3ccaf640 -1 osd.5 0 waiting for initial osdmap 2026-03-09T19:53:26.624 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:26 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:26.351+0000 7f3d38ad9640 -1 osd.5 33 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-09T19:53:26.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:26.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: Detected new or changed devices on vm11 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: Adjusting osd_memory_target on vm11 to 2176M 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: osdmap e32: 6 total, 5 up, 6 in 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: from='osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]: dispatch 2026-03-09T19:53:26.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:26 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2256813375' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:27.208 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd stat -f json 2026-03-09T19:53:27.383 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:27.495 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:27 vm02 ceph-mon[51633]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:27.496 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:27 vm02 ceph-mon[51633]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T19:53:27.496 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:27 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.496 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:27 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.610 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:27.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:27 vm11 ceph-mon[76771]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:27.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:27 vm11 ceph-mon[76771]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T19:53:27.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:27 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:27 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.690 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":34,"num_osds":6,"num_up_osds":6,"osd_up_since":1773086007,"num_in_osds":6,"osd_in_since":1773085997,"num_remapped_pgs":1} 2026-03-09T19:53:27.691 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd dump --format=json 2026-03-09T19:53:27.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:27 vm10 ceph-mon[55027]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm11", "root=default"]}]': finished 2026-03-09T19:53:27.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:27 vm10 ceph-mon[55027]: osdmap e33: 6 total, 5 up, 6 in 2026-03-09T19:53:27.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:27 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:27 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:27.871 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:28.103 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:28.103 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":34,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","created":"2026-03-09T19:51:34.063009+0000","modified":"2026-03-09T19:53:27.343287+0000","last_up_change":"2026-03-09T19:53:27.343287+0000","last_in_change":"2026-03-09T19:53:17.282268+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T19:52:58.798151+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5bf6a389-7f23-4b97-8b5b-9b8f72542309","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6803","nonce":1639950959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6805","nonce":1639950959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6809","nonce":1639950959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6807","nonce":1639950959}]},"public_addr":"192.168.123.102:6803/1639950959","cluster_addr":"192.168.123.102:6805/1639950959","heartbeat_back_addr":"192.168.123.102:6809/1639950959","heartbeat_front_addr":"192.168.123.102:6807/1639950959","state":["exists","up"]},{"osd":1,"uuid":"34047d7e-0449-4e8b-bb14-2f79ca4f35d2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6811","nonce":3659897092}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6813","nonce":3659897092}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6817","nonce":3659897092}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6815","nonce":3659897092}]},"public_addr":"192.168.123.102:6811/3659897092","cluster_addr":"192.168.123.102:6813/3659897092","heartbeat_back_addr":"192.168.123.102:6817/3659897092","heartbeat_front_addr":"192.168.123.102:6815/3659897092","state":["exists","up"]},{"osd":2,"uuid":"bd50b4cb-b42e-4544-8af8-f525f5e040d1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":18,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6801","nonce":1532484340}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6803","nonce":1532484340}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6807","nonce":1532484340}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6805","nonce":1532484340}]},"public_addr":"192.168.123.110:6801/1532484340","cluster_addr":"192.168.123.110:6803/1532484340","heartbeat_back_addr":"192.168.123.110:6807/1532484340","heartbeat_front_addr":"192.168.123.110:6805/1532484340","state":["exists","up"]},{"osd":3,"uuid":"133b4719-ab0d-4d74-a537-a82bc3bfb22f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":33,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6809","nonce":3577442477}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6811","nonce":3577442477}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6815","nonce":3577442477}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6813","nonce":3577442477}]},"public_addr":"192.168.123.110:6809/3577442477","cluster_addr":"192.168.123.110:6811/3577442477","heartbeat_back_addr":"192.168.123.110:6815/3577442477","heartbeat_front_addr":"192.168.123.110:6813/3577442477","state":["exists","up"]},{"osd":4,"uuid":"4bb90fd1-5e97-470b-b51e-8fefb71fa68d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":29,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6801","nonce":3542558984}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6803","nonce":3542558984}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6807","nonce":3542558984}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6805","nonce":3542558984}]},"public_addr":"192.168.123.111:6801/3542558984","cluster_addr":"192.168.123.111:6803/3542558984","heartbeat_back_addr":"192.168.123.111:6807/3542558984","heartbeat_front_addr":"192.168.123.111:6805/3542558984","state":["exists","up"]},{"osd":5,"uuid":"2b1993e3-b0a4-4651-a5df-7776ee00c4e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6809","nonce":2473834247}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6811","nonce":2473834247}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6815","nonce":2473834247}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6813","nonce":2473834247}]},"public_addr":"192.168.123.111:6809/2473834247","cluster_addr":"192.168.123.111:6811/2473834247","heartbeat_back_addr":"192.168.123.111:6815/2473834247","heartbeat_front_addr":"192.168.123.111:6813/2473834247","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:34.690254+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:45.578329+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:55.578733+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:05.422507+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:16.125856+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[{"pgid":"1.0","osds":[3,2,4]}],"primary_temp":[],"blocklist":{"192.168.123.102:0/863259861":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/1155817802":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/679339282":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/925503083":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/4071564622":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/3883138366":"2026-03-10T19:51:45.498055+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T19:53:28.155 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-09T19:52:58.798151+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '20', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 6, 'score_stable': 6, 'optimal_score': 0.5, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-09T19:53:28.155 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd pool get .mgr pg_num 2026-03-09T19:53:28.342 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: purged_snaps scrub starts 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247] boot 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: osdmap e34: 6 total, 6 up, 6 in 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/4129332867' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:28.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:28 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3340947173' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:28.592 INFO:teuthology.orchestra.run.vm02.stdout:pg_num: 1 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: purged_snaps scrub starts 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247] boot 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: osdmap e34: 6 total, 6 up, 6 in 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/4129332867' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:28.614 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:28 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3340947173' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: purged_snaps scrub starts 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: purged_snaps scrub ok 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 534 MiB used, 99 GiB / 100 GiB avail; 69 KiB/s, 0 objects/s recovering 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: osd.5 [v2:192.168.123.111:6808/2473834247,v1:192.168.123.111:6809/2473834247] boot 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: osdmap e34: 6 total, 6 up, 6 in 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/4129332867' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-09T19:53:28.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:28 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3340947173' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:28.690 INFO:tasks.cephadm:Setting up client nodes... 2026-03-09T19:53:28.690 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-09T19:53:28.690 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-09T19:53:28.690 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph mgr dump --format=json 2026-03-09T19:53:28.865 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:29.169 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:29.241 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"flags":0,"active_gid":14150,"active_name":"a","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":3913487111},{"type":"v1","addr":"192.168.123.102:6801","nonce":3913487111}]},"active_addr":"192.168.123.102:6801/3913487111","active_change":"2026-03-09T19:51:54.746473+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14211,"name":"b","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.102:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":1366177219}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":4022562082}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":1815850867}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":4185431618}]}]} 2026-03-09T19:53:29.243 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-09T19:53:29.243 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-09T19:53:29.243 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd dump --format=json 2026-03-09T19:53:29.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:29 vm02 ceph-mon[51633]: osdmap e35: 6 total, 6 up, 6 in 2026-03-09T19:53:29.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:29 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2760490079' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T19:53:29.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:29 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3895733263' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T19:53:29.427 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:29.622 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:29 vm10 ceph-mon[55027]: osdmap e35: 6 total, 6 up, 6 in 2026-03-09T19:53:29.622 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:29 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2760490079' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T19:53:29.622 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:29 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3895733263' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T19:53:29.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:29 vm11 ceph-mon[76771]: osdmap e35: 6 total, 6 up, 6 in 2026-03-09T19:53:29.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:29 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2760490079' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-09T19:53:29.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:29 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3895733263' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-09T19:53:29.664 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:29.672 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":36,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","created":"2026-03-09T19:51:34.063009+0000","modified":"2026-03-09T19:53:29.359687+0000","last_up_change":"2026-03-09T19:53:27.343287+0000","last_in_change":"2026-03-09T19:53:17.282268+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T19:52:58.798151+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5bf6a389-7f23-4b97-8b5b-9b8f72542309","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6803","nonce":1639950959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6805","nonce":1639950959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6809","nonce":1639950959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6807","nonce":1639950959}]},"public_addr":"192.168.123.102:6803/1639950959","cluster_addr":"192.168.123.102:6805/1639950959","heartbeat_back_addr":"192.168.123.102:6809/1639950959","heartbeat_front_addr":"192.168.123.102:6807/1639950959","state":["exists","up"]},{"osd":1,"uuid":"34047d7e-0449-4e8b-bb14-2f79ca4f35d2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6811","nonce":3659897092}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6813","nonce":3659897092}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6817","nonce":3659897092}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6815","nonce":3659897092}]},"public_addr":"192.168.123.102:6811/3659897092","cluster_addr":"192.168.123.102:6813/3659897092","heartbeat_back_addr":"192.168.123.102:6817/3659897092","heartbeat_front_addr":"192.168.123.102:6815/3659897092","state":["exists","up"]},{"osd":2,"uuid":"bd50b4cb-b42e-4544-8af8-f525f5e040d1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":18,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6801","nonce":1532484340}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6803","nonce":1532484340}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6807","nonce":1532484340}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6805","nonce":1532484340}]},"public_addr":"192.168.123.110:6801/1532484340","cluster_addr":"192.168.123.110:6803/1532484340","heartbeat_back_addr":"192.168.123.110:6807/1532484340","heartbeat_front_addr":"192.168.123.110:6805/1532484340","state":["exists","up"]},{"osd":3,"uuid":"133b4719-ab0d-4d74-a537-a82bc3bfb22f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":35,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6809","nonce":3577442477}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6811","nonce":3577442477}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6815","nonce":3577442477}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6813","nonce":3577442477}]},"public_addr":"192.168.123.110:6809/3577442477","cluster_addr":"192.168.123.110:6811/3577442477","heartbeat_back_addr":"192.168.123.110:6815/3577442477","heartbeat_front_addr":"192.168.123.110:6813/3577442477","state":["exists","up"]},{"osd":4,"uuid":"4bb90fd1-5e97-470b-b51e-8fefb71fa68d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":29,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6801","nonce":3542558984}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6803","nonce":3542558984}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6807","nonce":3542558984}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6805","nonce":3542558984}]},"public_addr":"192.168.123.111:6801/3542558984","cluster_addr":"192.168.123.111:6803/3542558984","heartbeat_back_addr":"192.168.123.111:6807/3542558984","heartbeat_front_addr":"192.168.123.111:6805/3542558984","state":["exists","up"]},{"osd":5,"uuid":"2b1993e3-b0a4-4651-a5df-7776ee00c4e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6809","nonce":2473834247}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6811","nonce":2473834247}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6815","nonce":2473834247}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6813","nonce":2473834247}]},"public_addr":"192.168.123.111:6809/2473834247","cluster_addr":"192.168.123.111:6811/2473834247","heartbeat_back_addr":"192.168.123.111:6815/2473834247","heartbeat_front_addr":"192.168.123.111:6813/2473834247","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:34.690254+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:45.578329+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:55.578733+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:05.422507+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:16.125856+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:26.102524+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/863259861":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/1155817802":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/679339282":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/925503083":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/4071564622":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/3883138366":"2026-03-10T19:51:45.498055+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T19:53:29.871 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-09T19:53:29.871 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd dump --format=json 2026-03-09T19:53:30.057 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:30.306 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:30.306 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":36,"fsid":"3527d82a-1bf1-11f1-92a4-57f58e2eb8dd","created":"2026-03-09T19:51:34.063009+0000","modified":"2026-03-09T19:53:29.359687+0000","last_up_change":"2026-03-09T19:53:27.343287+0000","last_in_change":"2026-03-09T19:53:17.282268+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":14,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":6,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-09T19:52:58.798151+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"20","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":6,"score_stable":6,"optimal_score":0.5,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5bf6a389-7f23-4b97-8b5b-9b8f72542309","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":8,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6803","nonce":1639950959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6805","nonce":1639950959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6809","nonce":1639950959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":1639950959},{"type":"v1","addr":"192.168.123.102:6807","nonce":1639950959}]},"public_addr":"192.168.123.102:6803/1639950959","cluster_addr":"192.168.123.102:6805/1639950959","heartbeat_back_addr":"192.168.123.102:6809/1639950959","heartbeat_front_addr":"192.168.123.102:6807/1639950959","state":["exists","up"]},{"osd":1,"uuid":"34047d7e-0449-4e8b-bb14-2f79ca4f35d2","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6811","nonce":3659897092}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6813","nonce":3659897092}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6817","nonce":3659897092}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":3659897092},{"type":"v1","addr":"192.168.123.102:6815","nonce":3659897092}]},"public_addr":"192.168.123.102:6811/3659897092","cluster_addr":"192.168.123.102:6813/3659897092","heartbeat_back_addr":"192.168.123.102:6817/3659897092","heartbeat_front_addr":"192.168.123.102:6815/3659897092","state":["exists","up"]},{"osd":2,"uuid":"bd50b4cb-b42e-4544-8af8-f525f5e040d1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":18,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6800","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6801","nonce":1532484340}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6802","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6803","nonce":1532484340}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6806","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6807","nonce":1532484340}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6804","nonce":1532484340},{"type":"v1","addr":"192.168.123.110:6805","nonce":1532484340}]},"public_addr":"192.168.123.110:6801/1532484340","cluster_addr":"192.168.123.110:6803/1532484340","heartbeat_back_addr":"192.168.123.110:6807/1532484340","heartbeat_front_addr":"192.168.123.110:6805/1532484340","state":["exists","up"]},{"osd":3,"uuid":"133b4719-ab0d-4d74-a537-a82bc3bfb22f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":35,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6808","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6809","nonce":3577442477}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6810","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6811","nonce":3577442477}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6814","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6815","nonce":3577442477}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.110:6812","nonce":3577442477},{"type":"v1","addr":"192.168.123.110:6813","nonce":3577442477}]},"public_addr":"192.168.123.110:6809/3577442477","cluster_addr":"192.168.123.110:6811/3577442477","heartbeat_back_addr":"192.168.123.110:6815/3577442477","heartbeat_front_addr":"192.168.123.110:6813/3577442477","state":["exists","up"]},{"osd":4,"uuid":"4bb90fd1-5e97-470b-b51e-8fefb71fa68d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":29,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6800","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6801","nonce":3542558984}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6802","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6803","nonce":3542558984}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6806","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6807","nonce":3542558984}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6804","nonce":3542558984},{"type":"v1","addr":"192.168.123.111:6805","nonce":3542558984}]},"public_addr":"192.168.123.111:6801/3542558984","cluster_addr":"192.168.123.111:6803/3542558984","heartbeat_back_addr":"192.168.123.111:6807/3542558984","heartbeat_front_addr":"192.168.123.111:6805/3542558984","state":["exists","up"]},{"osd":5,"uuid":"2b1993e3-b0a4-4651-a5df-7776ee00c4e5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":34,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6808","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6809","nonce":2473834247}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6810","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6811","nonce":2473834247}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6814","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6815","nonce":2473834247}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.111:6812","nonce":2473834247},{"type":"v1","addr":"192.168.123.111:6813","nonce":2473834247}]},"public_addr":"192.168.123.111:6809/2473834247","cluster_addr":"192.168.123.111:6811/2473834247","heartbeat_back_addr":"192.168.123.111:6815/2473834247","heartbeat_front_addr":"192.168.123.111:6813/2473834247","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:34.690254+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:45.578329+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:52:55.578733+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:05.422507+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:16.125856+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-09T19:53:26.102524+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/863259861":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/1155817802":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/679339282":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:6801/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/986462617":"2026-03-10T19:51:45.498055+0000","192.168.123.102:6800/3456911997":"2026-03-10T19:51:54.746368+0000","192.168.123.102:0/925503083":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/4071564622":"2026-03-10T19:51:45.498055+0000","192.168.123.102:0/3883138366":"2026-03-10T19:51:45.498055+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-09T19:53:30.381 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.0 flush_pg_stats 2026-03-09T19:53:30.381 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.1 flush_pg_stats 2026-03-09T19:53:30.381 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.2 flush_pg_stats 2026-03-09T19:53:30.382 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.3 flush_pg_stats 2026-03-09T19:53:30.382 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.4 flush_pg_stats 2026-03-09T19:53:30.382 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph tell osd.5 flush_pg_stats 2026-03-09T19:53:30.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:30 vm11 ceph-mon[76771]: pgmap v72: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:30.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:30 vm11 ceph-mon[76771]: osdmap e36: 6 total, 6 up, 6 in 2026-03-09T19:53:30.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:30 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/4110267758' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:30 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2002339990' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:30 vm02 ceph-mon[51633]: pgmap v72: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:30.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:30 vm02 ceph-mon[51633]: osdmap e36: 6 total, 6 up, 6 in 2026-03-09T19:53:30.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:30 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/4110267758' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:30 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2002339990' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:30 vm10 ceph-mon[55027]: pgmap v72: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:30.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:30 vm10 ceph-mon[55027]: osdmap e36: 6 total, 6 up, 6 in 2026-03-09T19:53:30.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:30 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/4110267758' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:30 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2002339990' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-09T19:53:30.909 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:30.917 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:30.941 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:30.958 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:30.958 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:31.379 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:31.763 INFO:teuthology.orchestra.run.vm02.stdout:73014444040 2026-03-09T19:53:31.764 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.2 2026-03-09T19:53:31.812 INFO:teuthology.orchestra.run.vm02.stdout:55834574858 2026-03-09T19:53:31.813 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.1 2026-03-09T19:53:31.917 INFO:teuthology.orchestra.run.vm02.stdout:124554051589 2026-03-09T19:53:31.917 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.4 2026-03-09T19:53:31.947 INFO:teuthology.orchestra.run.vm02.stdout:146028888066 2026-03-09T19:53:31.948 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.5 2026-03-09T19:53:31.950 INFO:teuthology.orchestra.run.vm02.stdout:34359738381 2026-03-09T19:53:31.950 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.0 2026-03-09T19:53:31.958 INFO:teuthology.orchestra.run.vm02.stdout:98784247814 2026-03-09T19:53:31.958 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.3 2026-03-09T19:53:32.233 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:32.431 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:32.514 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:32.632 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:32 vm02 ceph-mon[51633]: pgmap v74: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:32.632 INFO:teuthology.orchestra.run.vm02.stdout:73014444039 2026-03-09T19:53:32.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:32 vm10 ceph-mon[55027]: pgmap v74: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:32.713 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:32 vm11 ceph-mon[76771]: pgmap v74: 1 pgs: 1 peering; 449 KiB data, 160 MiB used, 120 GiB / 120 GiB avail 2026-03-09T19:53:32.761 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444040 got 73014444039 for osd.2 2026-03-09T19:53:32.826 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:32.833 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:32.990 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:33.085 INFO:teuthology.orchestra.run.vm02.stdout:34359738379 2026-03-09T19:53:33.138 INFO:teuthology.orchestra.run.vm02.stdout:55834574857 2026-03-09T19:53:33.217 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574858 got 55834574857 for osd.1 2026-03-09T19:53:33.263 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738381 got 34359738379 for osd.0 2026-03-09T19:53:33.352 INFO:teuthology.orchestra.run.vm02.stdout:124554051588 2026-03-09T19:53:33.379 INFO:teuthology.orchestra.run.vm02.stdout:146028888065 2026-03-09T19:53:33.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:33 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2487377266' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:33.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:33 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2495826536' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:33.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:33 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/466250088' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:33.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:33 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/493290408' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:33.406 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:33 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3481759028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:33.438 INFO:tasks.cephadm.ceph_manager.ceph:need seq 124554051589 got 124554051588 for osd.4 2026-03-09T19:53:33.452 INFO:tasks.cephadm.ceph_manager.ceph:need seq 146028888066 got 146028888065 for osd.5 2026-03-09T19:53:33.495 INFO:teuthology.orchestra.run.vm02.stdout:98784247815 2026-03-09T19:53:33.568 INFO:tasks.cephadm.ceph_manager.ceph:need seq 98784247814 got 98784247815 for osd.3 2026-03-09T19:53:33.568 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:33.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:33 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2487377266' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:33.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:33 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2495826536' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:33.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:33 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/466250088' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:33.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:33 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/493290408' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:33.624 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:33 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3481759028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:33.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:33 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2487377266' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:33.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:33 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2495826536' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:33.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:33 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/466250088' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:33.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:33 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/493290408' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:33.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:33 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3481759028' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:33.762 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.2 2026-03-09T19:53:33.948 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:34.187 INFO:teuthology.orchestra.run.vm02.stdout:73014444040 2026-03-09T19:53:34.217 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.1 2026-03-09T19:53:34.264 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.0 2026-03-09T19:53:34.272 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444040 got 73014444040 for osd.2 2026-03-09T19:53:34.272 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:34.399 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:34 vm02 ceph-mon[51633]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T19:53:34.399 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:34 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1126115597' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T19:53:34.399 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:34 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2145863153' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:34.428 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:34.438 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.4 2026-03-09T19:53:34.453 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph osd last-stat-seq osd.5 2026-03-09T19:53:34.594 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:34.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:34 vm10 ceph-mon[55027]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T19:53:34.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:34 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/1126115597' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T19:53:34.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:34 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2145863153' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:34.851 INFO:teuthology.orchestra.run.vm02.stdout:55834574858 2026-03-09T19:53:34.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:34 vm11 ceph-mon[76771]: pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-09T19:53:34.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:34 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/1126115597' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-09T19:53:34.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:34 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/2145863153' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-09T19:53:34.938 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:34.954 INFO:tasks.cephadm.ceph_manager.ceph:need seq 55834574858 got 55834574858 for osd.1 2026-03-09T19:53:34.954 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:35.010 INFO:teuthology.orchestra.run.vm02.stdout:34359738381 2026-03-09T19:53:35.103 INFO:tasks.cephadm.ceph_manager.ceph:need seq 34359738381 got 34359738381 for osd.0 2026-03-09T19:53:35.103 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:35.115 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:35.256 INFO:teuthology.orchestra.run.vm02.stdout:124554051589 2026-03-09T19:53:35.339 INFO:tasks.cephadm.ceph_manager.ceph:need seq 124554051589 got 124554051589 for osd.4 2026-03-09T19:53:35.340 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:35.412 INFO:teuthology.orchestra.run.vm02.stdout:146028888067 2026-03-09T19:53:35.473 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:35 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3418659122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:35.473 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:35 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/1126203541' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:35.473 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:35 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/3306626741' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:35.505 INFO:tasks.cephadm.ceph_manager.ceph:need seq 146028888066 got 146028888067 for osd.5 2026-03-09T19:53:35.505 DEBUG:teuthology.parallel:result is None 2026-03-09T19:53:35.505 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-09T19:53:35.505 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph pg dump --format=json 2026-03-09T19:53:35.685 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:35.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:35 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3418659122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:35.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:35 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/1126203541' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:35.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:35 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/3306626741' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:35 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3418659122' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-09T19:53:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:35 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/1126203541' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-09T19:53:35.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:35 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/3306626741' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-09T19:53:35.916 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:35.917 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-09T19:53:35.966 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":76,"stamp":"2026-03-09T19:53:34.775204+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":3,"kb":125804544,"kb_used":164380,"kb_used_data":3324,"kb_used_omap":9,"kb_used_meta":160886,"kb_avail":125640164,"statfs":{"total":128823853056,"available":128655527936,"internally_reserved":0,"allocated":3403776,"data_stored":2129754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9537,"internal_metadata":164747967},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"7.424112"},"pg_stats":[{"pgid":"1.0","version":"19'32","reported_seq":57,"reported_epoch":36,"state":"active+clean","last_fresh":"2026-03-09T19:53:29.820223+0000","last_change":"2026-03-09T19:53:29.820223+0000","last_active":"2026-03-09T19:53:29.820223+0000","last_peered":"2026-03-09T19:53:29.820223+0000","last_clean":"2026-03-09T19:53:29.820223+0000","last_became_active":"2026-03-09T19:53:29.415146+0000","last_became_peered":"2026-03-09T19:53:29.415146+0000","last_unstale":"2026-03-09T19:53:29.820223+0000","last_undegraded":"2026-03-09T19:53:29.820223+0000","last_fullsized":"2026-03-09T19:53:29.820223+0000","mapping_epoch":35,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":36,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T19:52:59.510488+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T19:52:59.510488+0000","last_clean_scrub_stamp":"2026-03-09T19:52:59.510488+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:02:53.894458+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2777088,"data_stored":2755680,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":6}],"osd_stats":[{"osd":5,"up_from":34,"seq":146028888067,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27620,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939804,"statfs":{"total":21470642176,"available":21442359296,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":29,"seq":124554051590,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27172,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940252,"statfs":{"total":21470642176,"available":21442818048,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":23,"seq":98784247815,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27616,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939808,"statfs":{"total":21470642176,"available":21442363392,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":17,"seq":73014444041,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27620,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939804,"statfs":{"total":21470642176,"available":21442359296,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574859,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27176,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940248,"statfs":{"total":21470642176,"available":21442813952,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738381,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27176,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940248,"statfs":{"total":21470642176,"available":21442813952,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T19:53:35.967 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph pg dump --format=json 2026-03-09T19:53:36.160 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:36.414 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:36.414 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-09T19:53:36.469 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":76,"stamp":"2026-03-09T19:53:34.775204+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":6,"num_per_pool_osds":6,"num_per_pool_omap_osds":3,"kb":125804544,"kb_used":164380,"kb_used_data":3324,"kb_used_omap":9,"kb_used_meta":160886,"kb_avail":125640164,"statfs":{"total":128823853056,"available":128655527936,"internally_reserved":0,"allocated":3403776,"data_stored":2129754,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":9537,"internal_metadata":164747967},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":2,"num_bytes_recovered":459280,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"7.424112"},"pg_stats":[{"pgid":"1.0","version":"19'32","reported_seq":57,"reported_epoch":36,"state":"active+clean","last_fresh":"2026-03-09T19:53:29.820223+0000","last_change":"2026-03-09T19:53:29.820223+0000","last_active":"2026-03-09T19:53:29.820223+0000","last_peered":"2026-03-09T19:53:29.820223+0000","last_clean":"2026-03-09T19:53:29.820223+0000","last_became_active":"2026-03-09T19:53:29.415146+0000","last_became_peered":"2026-03-09T19:53:29.415146+0000","last_unstale":"2026-03-09T19:53:29.820223+0000","last_undegraded":"2026-03-09T19:53:29.820223+0000","last_fullsized":"2026-03-09T19:53:29.820223+0000","mapping_epoch":35,"log_start":"0'0","ondisk_log_start":"0'0","created":18,"last_epoch_clean":36,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-09T19:52:59.510488+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-09T19:52:59.510488+0000","last_clean_scrub_stamp":"2026-03-09T19:52:59.510488+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-10T21:02:53.894458+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":6,"num_bytes_recovered":1377840,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2777088,"data_stored":2755680,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":6}],"osd_stats":[{"osd":5,"up_from":34,"seq":146028888067,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27620,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939804,"statfs":{"total":21470642176,"available":21442359296,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":29,"seq":124554051590,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27172,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940252,"statfs":{"total":21470642176,"available":21442818048,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":23,"seq":98784247815,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27616,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939808,"statfs":{"total":21470642176,"available":21442363392,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":17,"seq":73014444041,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27620,"kb_used_data":780,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939804,"statfs":{"total":21470642176,"available":21442359296,"internally_reserved":0,"allocated":798720,"data_stored":584599,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574859,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27176,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940248,"statfs":{"total":21470642176,"available":21442813952,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":8,"seq":34359738381,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27176,"kb_used_data":328,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940248,"statfs":{"total":21470642176,"available":21442813952,"internally_reserved":0,"allocated":335872,"data_stored":125319,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[1,2,3,4,5],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-09T19:53:36.469 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-09T19:53:36.469 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-09T19:53:36.469 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-09T19:53:36.470 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph health --format=json 2026-03-09T19:53:36.498 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:36 vm02 ceph-mon[51633]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 60 KiB/s, 0 objects/s recovering 2026-03-09T19:53:36.498 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:36 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/503434468' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:36.658 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:36.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:36 vm10 ceph-mon[55027]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 60 KiB/s, 0 objects/s recovering 2026-03-09T19:53:36.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:36 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/503434468' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:36 vm11 ceph-mon[76771]: pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 60 KiB/s, 0 objects/s recovering 2026-03-09T19:53:36.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:36 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/503434468' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-09T19:53:36.920 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:53:36.921 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-09T19:53:36.994 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-09T19:53:36.995 INFO:tasks.cephadm:Setup complete, yielding 2026-03-09T19:53:36.995 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-09T19:53:36.997 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm02.local 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- bash -c 'set -ex 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> HOSTNAMES=$(ceph orch host ls --format json | jq -r '"'"'.[] | .hostname'"'"') 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> for host in $HOSTNAMES; do 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # find the hostname for "host.c" which will have no mgr 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> HAS_MGRS=$(ceph orch ps --hostname ${host} --format json | jq '"'"'any(.daemon_type == "mgr")'"'"') 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> if [ "$HAS_MGRS" == "false" ]; then 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> HOST_C="${host}" 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> fi 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> done 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # One last thing to worry about before draining the host 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # is that the teuthology test tends to put the explicit 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # hostnames in the placement for the mon service. 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # We want to make sure we can drain without providing 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # --force and there is a check for the host being removed 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # being listed explicitly in the placements. Therefore, 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # we should remove it from the mon placement. 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph orch ls mon --export > mon.yaml 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> sed /"$HOST_C"/d mon.yaml > mon_adjusted.yaml 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph orch apply -i mon_adjusted.yaml 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # now drain that host 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph orch host drain $HOST_C --zap-osd-devices 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # wait for drain to complete 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C) 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> while [ "$HOST_C_DAEMONS" != "No daemons reported" ]; do 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> sleep 15 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> HOST_C_DAEMONS=$(ceph orch ps --hostname $HOST_C) 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> done 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # we want to check the ability to remove the host from 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # the CRUSH map, so we should first verify the host is in 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # the CRUSH map. 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph osd getcrushmap -o compiled-crushmap 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> crushtool -d compiled-crushmap -o crushmap.txt 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> CRUSH_MAP=$(cat crushmap.txt) 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> if ! grep -q "$HOST_C" <<< "$CRUSH_MAP"; then 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> printf "Expected to see $HOST_C in CRUSH map. Saw:\n\n$CRUSH_MAP" 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> exit 1 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> fi 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # If the drain was successful, we should be able to remove the 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # host without force with no issues. If there are still daemons 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # we will get a response telling us to drain the host and a 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # non-zero return code 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph orch host rm $HOST_C --rm-crush-entry 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> # verify we'"'"'ve successfully removed the host from the CRUSH map 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> sleep 30 2026-03-09T19:53:36.997 DEBUG:teuthology.orchestra.run.vm02:> ceph osd getcrushmap -o compiled-crushmap 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> crushtool -d compiled-crushmap -o crushmap.txt 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> CRUSH_MAP=$(cat crushmap.txt) 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> if grep -q "$HOST_C" <<< "$CRUSH_MAP"; then 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> printf "Saw $HOST_C in CRUSH map after it should have been removed.\n\n$CRUSH_MAP" 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> exit 1 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> fi 2026-03-09T19:53:36.998 DEBUG:teuthology.orchestra.run.vm02:> ' 2026-03-09T19:53:37.224 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:53:37.300 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch host ls --format json 2026-03-09T19:53:37.303 INFO:teuthology.orchestra.run.vm02.stderr:++ jq -r '.[] | .hostname' 2026-03-09T19:53:37.478 INFO:teuthology.orchestra.run.vm02.stderr:+ HOSTNAMES='vm02 2026-03-09T19:53:37.479 INFO:teuthology.orchestra.run.vm02.stderr:vm10 2026-03-09T19:53:37.479 INFO:teuthology.orchestra.run.vm02.stderr:vm11' 2026-03-09T19:53:37.479 INFO:teuthology.orchestra.run.vm02.stderr:+ for host in $HOSTNAMES 2026-03-09T19:53:37.479 INFO:teuthology.orchestra.run.vm02.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-09T19:53:37.479 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch ps --hostname vm02 --format json 2026-03-09T19:53:37.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:37 vm02 ceph-mon[51633]: from='client.14511 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:37 vm02 ceph-mon[51633]: from='client.14517 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.653 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:37 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/4008789145' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T19:53:37.653 INFO:teuthology.orchestra.run.vm02.stderr:+ HAS_MGRS=true 2026-03-09T19:53:37.653 INFO:teuthology.orchestra.run.vm02.stderr:+ '[' true == false ']' 2026-03-09T19:53:37.653 INFO:teuthology.orchestra.run.vm02.stderr:+ for host in $HOSTNAMES 2026-03-09T19:53:37.653 INFO:teuthology.orchestra.run.vm02.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-09T19:53:37.654 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch ps --hostname vm10 --format json 2026-03-09T19:53:37.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:37 vm10 ceph-mon[55027]: from='client.14511 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:37 vm10 ceph-mon[55027]: from='client.14517 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:37 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/4008789145' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T19:53:37.836 INFO:teuthology.orchestra.run.vm02.stderr:+ HAS_MGRS=true 2026-03-09T19:53:37.836 INFO:teuthology.orchestra.run.vm02.stderr:+ '[' true == false ']' 2026-03-09T19:53:37.836 INFO:teuthology.orchestra.run.vm02.stderr:+ for host in $HOSTNAMES 2026-03-09T19:53:37.836 INFO:teuthology.orchestra.run.vm02.stderr:++ jq 'any(.daemon_type == "mgr")' 2026-03-09T19:53:37.837 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch ps --hostname vm11 --format json 2026-03-09T19:53:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:37 vm11 ceph-mon[76771]: from='client.14511 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:37 vm11 ceph-mon[76771]: from='client.14517 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:37.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:37 vm11 ceph-mon[76771]: from='client.? 192.168.123.102:0/4008789145' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-09T19:53:38.020 INFO:teuthology.orchestra.run.vm02.stderr:+ HAS_MGRS=false 2026-03-09T19:53:38.020 INFO:teuthology.orchestra.run.vm02.stderr:+ '[' false == false ']' 2026-03-09T19:53:38.020 INFO:teuthology.orchestra.run.vm02.stderr:+ HOST_C=vm11 2026-03-09T19:53:38.020 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch ls mon --export 2026-03-09T19:53:38.203 INFO:teuthology.orchestra.run.vm02.stderr:+ sed /vm11/d mon.yaml 2026-03-09T19:53:38.204 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch apply -i mon_adjusted.yaml 2026-03-09T19:53:38.375 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled mon update... 2026-03-09T19:53:38.394 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch host drain vm11 --zap-osd-devices 2026-03-09T19:53:38.874 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:38 vm11 podman[87994]: 2026-03-09 19:53:38.42630986 +0000 UTC m=+0.012970830 container died bff86f9139506823bbbece1e55b24352898210a86b2441a33f956d7d23df9026 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-c, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, io.buildah.version=1.41.3, OSD_FLAVOR=default, org.label-schema.vendor=CentOS) 2026-03-09T19:53:38.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:38 vm11 podman[87994]: 2026-03-09 19:53:38.442275223 +0000 UTC m=+0.028936204 container remove bff86f9139506823bbbece1e55b24352898210a86b2441a33f956d7d23df9026 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-c, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, CEPH_REF=squid, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223) 2026-03-09T19:53:38.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:38 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.c.service: Deactivated successfully. 2026-03-09T19:53:38.875 INFO:journalctl@ceph.mon.c.vm11.stdout:Mar 09 19:53:38 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.c.service: Consumed 1.413s CPU time. 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: purged_snaps scrub ok 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='client.14535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm02", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='client.14541 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm10", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='client.14547 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='client.14553 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_type": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='client.24386 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: Saving service mon spec with placement vm02:192.168.123.102=a;vm10:192.168.123.110=b;count:3 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: Safe to remove mon.c: new quorum should be ['a', 'b'] (from ['a', 'b']) 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: Removing monitor c from monmap... 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "mon rm", "name": "c"}]': finished 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: mon.b calling monitor election 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: mon.a calling monitor election 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: monmap epoch 4 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: last_changed 2026-03-09T19:53:38.388071+0000 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: min_mon_release 19 (squid) 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: election_strategy: 1 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: fsmap 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: osdmap e36: 6 total, 6 up, 6 in 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: mgrmap e14: a(active, since 105s), standbys: b 2026-03-09T19:53:40.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:40 vm10 ceph-mon[55027]: overall HEALTH_OK 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: purged_snaps scrub ok 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='client.14535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm02", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='client.14541 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm10", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='client.14547 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='client.14553 -' entity='client.admin' cmd=[{"prefix": "orch ls", "service_type": "mon", "export": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='client.24386 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: Saving service mon spec with placement vm02:192.168.123.102=a;vm10:192.168.123.110=b;count:3 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: Safe to remove mon.c: new quorum should be ['a', 'b'] (from ['a', 'b']) 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: Removing monitor c from monmap... 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "mon rm", "name": "c"}]': finished 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: mon.b calling monitor election 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "a"}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mon metadata", "id": "b"}]: dispatch 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: mon.a calling monitor election 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: mon.a is new leader, mons a,b in quorum (ranks 0,1) 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: monmap epoch 4 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: last_changed 2026-03-09T19:53:38.388071+0000 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: created 2026-03-09T19:51:32.974864+0000 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: min_mon_release 19 (squid) 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: election_strategy: 1 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: 1: [v2:192.168.123.110:3300/0,v1:192.168.123.110:6789/0] mon.b 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: fsmap 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: osdmap e36: 6 total, 6 up, 6 in 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: mgrmap e14: a(active, since 105s), standbys: b 2026-03-09T19:53:40.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:40 vm02 ceph-mon[51633]: overall HEALTH_OK 2026-03-09T19:53:41.602 INFO:teuthology.orchestra.run.vm02.stdout:Scheduled to remove the following daemons from host 'vm11' 2026-03-09T19:53:41.602 INFO:teuthology.orchestra.run.vm02.stdout:type id 2026-03-09T19:53:41.602 INFO:teuthology.orchestra.run.vm02.stdout:-------------------- --------------- 2026-03-09T19:53:41.602 INFO:teuthology.orchestra.run.vm02.stdout:osd 4 2026-03-09T19:53:41.602 INFO:teuthology.orchestra.run.vm02.stdout:osd 5 2026-03-09T19:53:41.618 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch ps --hostname vm11 2026-03-09T19:53:41.814 INFO:teuthology.orchestra.run.vm02.stderr:+ HOST_C_DAEMONS='NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:osd.4 vm11 running (28s) 0s ago 28s 56.0M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 29bff6be3ce2 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:osd.5 vm11 running (18s) 0s ago 18s 34.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 5070a057187e ' 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:+ '[' 'NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:osd.4 vm11 running (28s) 0s ago 28s 56.0M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 29bff6be3ce2 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:osd.5 vm11 running (18s) 0s ago 18s 34.5M 2176M 19.2.3-678-ge911bdeb 654f31e6858e 5070a057187e ' '!=' 'No daemons reported' ']' 2026-03-09T19:53:41.815 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 15 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: Removing daemon mon.c from vm11 -- ports [] 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:41.815 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:41 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: Removing daemon mon.c from vm11 -- ports [] 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:41.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:41 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 39 KiB/s, 0 objects/s recovering 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm10:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: Updating vm02:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch host drain", "hostname": "vm11", "zap_osd_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-09T19:53:42.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]: dispatch 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 39 KiB/s, 0 objects/s recovering 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm02:/etc/ceph/ceph.conf 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm10:/etc/ceph/ceph.conf 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm11:/etc/ceph/ceph.conf 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm10:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: Updating vm02:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch host drain", "hostname": "vm11", "zap_osd_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush tree", "format": "json"}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-09T19:53:42.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:42 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]: dispatch 2026-03-09T19:53:43.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:42 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-a[51629]: 2026-03-09T19:53:42.887+0000 7f45934f4640 -1 mon.a@0(leader).osd e37 definitely_dead 0 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: Added label _no_schedule to host vm11 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: Added label _no_conf_keyring to host vm11 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: osd.4 crush weight is 0.0194854736328125 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: osd.5 crush weight is 0.0194854736328125 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]': finished 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: osdmap e37: 6 total, 6 up, 6 in 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-09T19:53:43.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:43 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["4"]}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: Added label _no_schedule to host vm11 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: Added label _no_conf_keyring to host vm11 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: osd.4 crush weight is 0.0194854736328125 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: osd.5 crush weight is 0.0194854736328125 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd ok-to-stop", "ids": ["5"]}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd crush reweight", "name": "osd.5", "weight": 0.0}]': finished 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: osdmap e37: 6 total, 6 up, 6 in 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-09T19:53:43.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:43 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["4"]}]: dispatch 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: Stopping Ceph osd.4 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:44.116+0000 7f338e535640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:44.116+0000 7f338e535640 -1 osd.4 36 *** Got signal Terminated *** 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4[80306]: 2026-03-09T19:53:44.116+0000 7f338e535640 -1 osd.4 36 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88712]: 2026-03-09 19:53:44.139424489 +0000 UTC m=+0.037408317 container died 29bff6be3ce2237c30887678eaf5503dcbb9b682c7378e65a510da45999f6d70 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, CEPH_REF=squid, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88712]: 2026-03-09 19:53:44.167267117 +0000 UTC m=+0.065250945 container remove 29bff6be3ce2237c30887678eaf5503dcbb9b682c7378e65a510da45999f6d70 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T19:53:44.314 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 bash[88712]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.314189658 +0000 UTC m=+0.017957502 container create 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.353438147 +0000 UTC m=+0.057205991 container init 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.359243659 +0000 UTC m=+0.063011503 container start 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.label-schema.license=GPLv2, CEPH_REF=squid, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.360445007 +0000 UTC m=+0.064212842 container attach 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_REF=squid, org.label-schema.schema-version=1.0) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.306803728 +0000 UTC m=+0.010571572 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.494311649 +0000 UTC m=+0.198079484 container died 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.license=GPLv2, org.label-schema.build-date=20260223, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 podman[88781]: 2026-03-09 19:53:44.538745481 +0000 UTC m=+0.242513325 container remove 0412ad81a40bba5a541cd88404aadb5c9b28a8c14822de80e39dbb01d558b7bb (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-4-deactivate, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0) 2026-03-09T19:53:44.565 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service: Deactivated successfully. 2026-03-09T19:53:44.566 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service: Unit process 88793 (conmon) remains running after unit stopped. 2026-03-09T19:53:44.566 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service: Unit process 88801 (podman) remains running after unit stopped. 2026-03-09T19:53:44.566 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: Stopped Ceph osd.4 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:53:44.566 INFO:journalctl@ceph.osd.4.vm11.stdout:Mar 09 19:53:44 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service: Consumed 1.041s CPU time, 86.9M memory peak. 2026-03-09T19:53:44.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: osd.5 weight is now 0.0 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["4"]}]': finished 2026-03-09T19:53:44.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:44 vm02 ceph-mon[51633]: osdmap e38: 6 total, 5 up, 6 in 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: osd.5 weight is now 0.0 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["4"]}]: dispatch 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["4"]}]': finished 2026-03-09T19:53:44.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:44 vm10 ceph-mon[55027]: osdmap e38: 6 total, 5 up, 6 in 2026-03-09T19:53:45.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: osd.4 now down 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: Removing daemon osd.4 from vm11 -- ports [] 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.4"}]: dispatch 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.4"}]': finished 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: Cluster is now healthy 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: osdmap e39: 5 total, 5 up, 5 in 2026-03-09T19:53:45.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:45 vm02 ceph-mon[51633]: osdmap e40: 5 total, 5 up, 5 in 2026-03-09T19:53:45.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: osd.4 now down 2026-03-09T19:53:45.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: Removing daemon osd.4 from vm11 -- ports [] 2026-03-09T19:53:45.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:45.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.4"}]: dispatch 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.4"}]': finished 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: Cluster is now healthy 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 4, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: osdmap e39: 5 total, 5 up, 5 in 2026-03-09T19:53:45.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:45 vm10 ceph-mon[55027]: osdmap e40: 5 total, 5 up, 5 in 2026-03-09T19:53:46.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: Removing key for osd.4 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: Successfully removed osd.4 on vm11 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: Successfully purged osd.4 on vm11 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: Zapping devices for osd.4 on vm11 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: osdmap e41: 5 total, 5 up, 5 in 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:46.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:46 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:46.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: Removing key for osd.4 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: Successfully removed osd.4 on vm11 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: Successfully purged osd.4 on vm11 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: Zapping devices for osd.4 on vm11 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: osdmap e41: 5 total, 5 up, 5 in 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:46.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:46 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: Successfully zapped devices for osd.4 on vm11 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:53:47.900 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:47 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:47.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: Successfully zapped devices for osd.4 on vm11 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:53:47.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:47 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Detected new or changed devices on vm11 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Removing vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Removing vm11:/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Removing vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.client.admin.keyring 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Reconfiguring mon.a (monmap changed)... 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: Reconfiguring daemon mon.a on vm02 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T19:53:48.791 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:48 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Detected new or changed devices on vm11 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Removing vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.conf 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Removing vm11:/etc/ceph/ceph.client.admin.keyring 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Removing vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/config/ceph.client.admin.keyring 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Reconfiguring mon.a (monmap changed)... 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: Reconfiguring daemon mon.a on vm02 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.a", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-09T19:53:48.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:48 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring daemon mgr.a on vm02 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring daemon osd.0 on vm02 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring daemon osd.1 on vm02 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring mon.b (monmap changed)... 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: Reconfiguring daemon mon.b on vm10 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:53:49.577 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T19:53:49.578 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:49 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring mgr.a (monmap changed)... 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring daemon mgr.a on vm02 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring osd.0 (monmap changed)... 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring daemon osd.0 on vm02 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring osd.1 (monmap changed)... 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring daemon osd.1 on vm02 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring mon.b (monmap changed)... 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: Reconfiguring daemon mon.b on vm10 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.b", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-09T19:53:49.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:49 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:50.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: pgmap v88: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 91 KiB/s, 0 objects/s recovering 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring daemon mgr.b on vm10 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring daemon osd.2 on vm10 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring osd.3 (monmap changed)... 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: Reconfiguring daemon osd.3 on vm10 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:50 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: pgmap v88: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 91 KiB/s, 0 objects/s recovering 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring mgr.b (monmap changed)... 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring daemon mgr.b on vm10 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring osd.2 (monmap changed)... 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring daemon osd.2 on vm10 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring osd.3 (monmap changed)... 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: Reconfiguring daemon osd.3 on vm10 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:50.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:50 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:51.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-a[51629]: 2026-03-09T19:53:51.118+0000 7f45934f4640 -1 mon.a@0(leader).osd e41 definitely_dead 0 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: Reconfiguring osd.5 (monmap changed)... 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: Reconfiguring daemon osd.5 on vm11 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-09T19:53:51.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:51 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["5"]}]: dispatch 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: Reconfiguring osd.5 (monmap changed)... 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: Reconfiguring daemon osd.5 on vm11 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:51.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:51.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-09T19:53:51.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:51 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd down", "ids": ["5"]}]: dispatch 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 systemd[1]: Stopping Ceph osd.5 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:52.354+0000 7f3d3dcc3640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:52.354+0000 7f3d3dcc3640 -1 osd.5 41 *** Got signal Terminated *** 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5[85186]: 2026-03-09T19:53:52.354+0000 7f3d3dcc3640 -1 osd.5 41 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93572]: 2026-03-09 19:53:52.401397881 +0000 UTC m=+0.061979653 container died 5070a057187ec7facb515f366810cd3c6a2bf0056244900a651dec87e54d6069 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS) 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93572]: 2026-03-09 19:53:52.416016999 +0000 UTC m=+0.076598760 container remove 5070a057187ec7facb515f366810cd3c6a2bf0056244900a651dec87e54d6069 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, ceph=True, io.buildah.version=1.41.3) 2026-03-09T19:53:52.562 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 bash[93572]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.56241925 +0000 UTC m=+0.017399082 container create 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, OSD_FLAVOR=default, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.595924743 +0000 UTC m=+0.050904576 container init 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, OSD_FLAVOR=default, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.600752339 +0000 UTC m=+0.055732171 container start 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, OSD_FLAVOR=default, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.605089195 +0000 UTC m=+0.060069027 container attach 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, OSD_FLAVOR=default) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.555925077 +0000 UTC m=+0.010904909 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 conmon[93652]: conmon 79c3856f598155070e9c : Failed to open cgroups file: /sys/fs/cgroup/machine.slice/libpod-79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a.scope/container/memory.events 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.73458145 +0000 UTC m=+0.189561282 container died 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 podman[93640]: 2026-03-09 19:53:52.747926224 +0000 UTC m=+0.202906045 container remove 79c3856f598155070e9cf2f122eb3b5230bbac4399649d2194da9c7d97deee7a (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-5-deactivate, OSD_FLAVOR=default, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, io.buildah.version=1.41.3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5.service: Deactivated successfully. 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5.service: Unit process 93652 (conmon) remains running after unit stopped. 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5.service: Unit process 93660 (podman) remains running after unit stopped. 2026-03-09T19:53:52.840 INFO:journalctl@ceph.osd.5.vm11.stdout:Mar 09 19:53:52 vm11 systemd[1]: Stopped Ceph osd.5 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: pgmap v89: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 74 KiB/s, 0 objects/s recovering 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["5"]}]': finished 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: osdmap e42: 5 total, 4 up, 5 in 2026-03-09T19:53:52.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: osd.5 now down 2026-03-09T19:53:52.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:52 vm02 ceph-mon[51633]: Removing daemon osd.5 from vm11 -- ports [] 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: pgmap v89: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 74 KiB/s, 0 objects/s recovering 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd df", "format": "json"}]: dispatch 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "osd safe-to-destroy", "ids": ["5"]}]: dispatch 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: Health check failed: 1 osds down (OSD_DOWN) 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-09T19:53:52.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd down", "ids": ["5"]}]': finished 2026-03-09T19:53:52.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: osdmap e42: 5 total, 4 up, 5 in 2026-03-09T19:53:52.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: osd.5 now down 2026-03-09T19:53:52.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:52 vm10 ceph-mon[55027]: Removing daemon osd.5 from vm11 -- ports [] 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: pgmap v91: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 60 KiB/s, 0 objects/s recovering 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Removing key for osd.5 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.5"}]': finished 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Successfully removed osd.5 on vm11 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Cluster is now healthy 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: osdmap e43: 4 total, 4 up, 4 in 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Successfully purged osd.5 on vm11 2026-03-09T19:53:53.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:53 vm02 ceph-mon[51633]: Zapping devices for osd.5 on vm11 2026-03-09T19:53:53.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: pgmap v91: 1 pgs: 1 active+clean+wait; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 60 KiB/s, 0 objects/s recovering 2026-03-09T19:53:53.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Removing key for osd.5 2026-03-09T19:53:53.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth rm", "entity": "osd.5"}]: dispatch 2026-03-09T19:53:53.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "auth rm", "entity": "osd.5"}]': finished 2026-03-09T19:53:53.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Successfully removed osd.5 on vm11 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]: dispatch 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Cluster is now healthy 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd purge-actual", "id": 5, "yes_i_really_mean_it": true}]': finished 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: osdmap e43: 4 total, 4 up, 4 in 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Successfully purged osd.5 on vm11 2026-03-09T19:53:53.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:53 vm10 ceph-mon[55027]: Zapping devices for osd.5 on vm11 2026-03-09T19:53:55.171 INFO:journalctl@ceph.mgr.a.vm02.stdout:Mar 09 19:53:54 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mgr-a[51844]: 2026-03-09T19:53:54.765+0000 7f9c77776640 -1 calc_pg_upmaps abort due to max <= 0 2026-03-09T19:53:55.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:55 vm02 ceph-mon[51633]: Successfully zapped devices for osd.5 on vm11 2026-03-09T19:53:55.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:55.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:55.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:55.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:55 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:55.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:55 vm10 ceph-mon[55027]: Successfully zapped devices for osd.5 on vm11 2026-03-09T19:53:55.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:55.712 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:55.713 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:55 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: pgmap v93: 1 pgs: 1 active+clean+wait; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.817 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:56 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:56.817 INFO:teuthology.orchestra.run.vm02.stderr:++ ceph orch ps --hostname vm11 2026-03-09T19:53:56.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: pgmap v93: 1 pgs: 1 active+clean+wait; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:56.963 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:56 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:56.989 INFO:teuthology.orchestra.run.vm02.stderr:+ HOST_C_DAEMONS='No daemons reported' 2026-03-09T19:53:56.990 INFO:teuthology.orchestra.run.vm02.stderr:+ '[' 'No daemons reported' '!=' 'No daemons reported' ']' 2026-03-09T19:53:56.990 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph osd getcrushmap -o compiled-crushmap 2026-03-09T19:53:57.155 INFO:teuthology.orchestra.run.vm02.stderr:17 2026-03-09T19:53:57.165 INFO:teuthology.orchestra.run.vm02.stderr:+ crushtool -d compiled-crushmap -o crushmap.txt 2026-03-09T19:53:57.175 INFO:teuthology.orchestra.run.vm02.stderr:++ cat crushmap.txt 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:+ CRUSH_MAP='# begin crush map 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_local_tries 0 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_local_fallback_tries 0 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_total_tries 50 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_descend_once 1 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_vary_r 1 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_stable 1 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable straw_calc_version 1 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:tunable allowed_bucket_algs 54 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:# devices 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:device 0 osd.0 class hdd 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:device 1 osd.1 class hdd 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:device 2 osd.2 class hdd 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:device 3 osd.3 class hdd 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:# types 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 0 osd 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 1 host 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 2 chassis 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 3 rack 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 4 row 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 5 pdu 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 6 pod 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 7 room 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 8 datacenter 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 9 zone 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 10 region 2026-03-09T19:53:57.178 INFO:teuthology.orchestra.run.vm02.stderr:type 11 root 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:# buckets 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:host vm02 { 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -3 # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -4 class hdd # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.03897 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item osd.0 weight 0.01949 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item osd.1 weight 0.01949 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:host vm10 { 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -5 # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -6 class hdd # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.03897 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item osd.2 weight 0.01949 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item osd.3 weight 0.01949 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:host vm11 { 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -7 # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -8 class hdd # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.00000 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:root default { 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -1 # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id -2 class hdd # do not change unnecessarily 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.07794 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item vm02 weight 0.03897 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item vm10 weight 0.03897 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: item vm11 weight 0.00000 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:# rules 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr:rule replicated_rule { 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: id 0 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: type replicated 2026-03-09T19:53:57.179 INFO:teuthology.orchestra.run.vm02.stderr: step take default 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr: step choose firstn 0 type osd 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr: step emit 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr:# end crush map' 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr:+ grep -q vm11 2026-03-09T19:53:57.180 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph orch host rm vm11 --rm-crush-entry 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: Detected new or changed devices on vm11 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2417769283' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:57 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush remove", "name": "vm11"}]: dispatch 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: Detected new or changed devices on vm11 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2417769283' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:57.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:57 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "osd crush remove", "name": "vm11"}]: dispatch 2026-03-09T19:53:58.269 INFO:teuthology.orchestra.run.vm02.stdout:Removed host 'vm11' 2026-03-09T19:53:58.279 INFO:teuthology.orchestra.run.vm02.stderr:+ sleep 30 2026-03-09T19:53:58.671 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: pgmap v94: 1 pgs: 1 active+clean+wait; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='client.14570 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='client.14578 -' entity='client.admin' cmd=[{"prefix": "orch host rm", "hostname": "vm11", "rm_crush_entry": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd crush remove", "name": "vm11"}]': finished 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: osdmap e44: 4 total, 4 up, 4 in 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix":"config-key del","key":"mgr/cephadm/host.vm11"}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.vm11"}]': finished 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:58.672 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:58 vm02 ceph-mon[51633]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: pgmap v94: 1 pgs: 1 active+clean+wait; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='client.14570 -' entity='client.admin' cmd=[{"prefix": "orch ps", "hostname": "vm11", "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='client.14578 -' entity='client.admin' cmd=[{"prefix": "orch host rm", "hostname": "vm11", "rm_crush_entry": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix": "osd crush remove", "name": "vm11"}]': finished 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: osdmap e44: 4 total, 4 up, 4 in 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix":"config-key del","key":"mgr/cephadm/host.vm11"}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd='[{"prefix":"config-key del","key":"mgr/cephadm/host.vm11"}]': finished 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-09T19:53:58.865 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:58 vm10 ceph-mon[55027]: from='mgr.14150 192.168.123.102:0/263411054' entity='mgr.a' 2026-03-09T19:53:59.670 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:53:59 vm02 ceph-mon[51633]: Removed host vm11 2026-03-09T19:53:59.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:53:59 vm10 ceph-mon[55027]: Removed host vm11 2026-03-09T19:54:00.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:00 vm02 ceph-mon[51633]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:00.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:00 vm10 ceph-mon[55027]: pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:02.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:02 vm02 ceph-mon[51633]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:02.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:02 vm10 ceph-mon[55027]: pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:04.921 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:04 vm02 ceph-mon[51633]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:04.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:04 vm10 ceph-mon[55027]: pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:05.920 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:05 vm02 ceph-mon[51633]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:05.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:05 vm10 ceph-mon[55027]: pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:08.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:07 vm02 ceph-mon[51633]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:08.212 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:07 vm10 ceph-mon[55027]: pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:09.962 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:09 vm10 ceph-mon[55027]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:10.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:09 vm02 ceph-mon[51633]: pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:12.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:11 vm02 ceph-mon[51633]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:12.212 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:11 vm10 ceph-mon[55027]: pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:14.212 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:13 vm10 ceph-mon[55027]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:14.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:13 vm02 ceph-mon[51633]: pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:16.212 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:15 vm10 ceph-mon[55027]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:16.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:15 vm02 ceph-mon[51633]: pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:18.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:17 vm02 ceph-mon[51633]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:18.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:17 vm10 ceph-mon[55027]: pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:20.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:19 vm02 ceph-mon[51633]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:20.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:19 vm10 ceph-mon[55027]: pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:22.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:22 vm02 ceph-mon[51633]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:22.463 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:22 vm10 ceph-mon[55027]: pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:24.421 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:24 vm02 ceph-mon[51633]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:24.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:24 vm10 ceph-mon[55027]: pgmap v108: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:26.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:26 vm02 ceph-mon[51633]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:26.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:26 vm10 ceph-mon[55027]: pgmap v109: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:28.280 INFO:teuthology.orchestra.run.vm02.stderr:+ ceph osd getcrushmap -o compiled-crushmap 2026-03-09T19:54:28.420 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:28 vm02 ceph-mon[51633]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:28.435 INFO:teuthology.orchestra.run.vm02.stderr:18 2026-03-09T19:54:28.445 INFO:teuthology.orchestra.run.vm02.stderr:+ crushtool -d compiled-crushmap -o crushmap.txt 2026-03-09T19:54:28.456 INFO:teuthology.orchestra.run.vm02.stderr:++ cat crushmap.txt 2026-03-09T19:54:28.459 INFO:teuthology.orchestra.run.vm02.stderr:+ CRUSH_MAP='# begin crush map 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_local_tries 0 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_local_fallback_tries 0 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable choose_total_tries 50 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_descend_once 1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_vary_r 1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable chooseleaf_stable 1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable straw_calc_version 1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:tunable allowed_bucket_algs 54 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:# devices 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:device 0 osd.0 class hdd 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:device 1 osd.1 class hdd 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:device 2 osd.2 class hdd 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:device 3 osd.3 class hdd 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:# types 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 0 osd 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 1 host 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 2 chassis 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 3 rack 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 4 row 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 5 pdu 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 6 pod 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 7 room 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 8 datacenter 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 9 zone 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 10 region 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:type 11 root 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:# buckets 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:host vm02 { 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -3 # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -4 class hdd # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.03897 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item osd.0 weight 0.01949 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item osd.1 weight 0.01949 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:host vm10 { 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -5 # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -6 class hdd # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.03897 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item osd.2 weight 0.01949 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item osd.3 weight 0.01949 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:root default { 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -1 # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: id -2 class hdd # do not change unnecessarily 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: # weight 0.07794 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: alg straw2 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: hash 0 # rjenkins1 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item vm02 weight 0.03897 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: item vm10 weight 0.03897 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:# rules 2026-03-09T19:54:28.460 INFO:teuthology.orchestra.run.vm02.stderr:rule replicated_rule { 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: id 0 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: type replicated 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: step take default 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: step choose firstn 0 type osd 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: step emit 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr:} 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr:# end crush map' 2026-03-09T19:54:28.461 INFO:teuthology.orchestra.run.vm02.stderr:+ grep -q vm11 2026-03-09T19:54:28.462 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:28 vm10 ceph-mon[55027]: pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-09T19:54:28.551 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-09T19:54:28.553 INFO:tasks.cephadm:Teardown begin 2026-03-09T19:54:28.553 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:28.580 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:28.609 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:28.641 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-09T19:54:28.641 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd -- ceph mgr module disable cephadm 2026-03-09T19:54:28.824 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/mon.a/config 2026-03-09T19:54:28.844 INFO:teuthology.orchestra.run.vm02.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-09T19:54:28.868 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-09T19:54:28.869 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-09T19:54:28.869 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T19:54:28.888 DEBUG:teuthology.orchestra.run.vm10:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T19:54:28.906 DEBUG:teuthology.orchestra.run.vm11:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-09T19:54:28.922 INFO:tasks.cephadm:Stopping all daemons... 2026-03-09T19:54:28.922 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-09T19:54:28.922 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a 2026-03-09T19:54:29.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:28 vm02 systemd[1]: Stopping Ceph mon.a for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:29.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:29 vm02 ceph-mon[51633]: from='client.? 192.168.123.102:0/2463944290' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-09T19:54:29.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:29 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-a[51629]: 2026-03-09T19:54:29.051+0000 7f4598cff640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:29.171 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 09 19:54:29 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-a[51629]: 2026-03-09T19:54:29.051+0000 7f4598cff640 -1 mon.a@0(leader) e4 *** Got Signal Terminated *** 2026-03-09T19:54:29.329 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.a.service' 2026-03-09T19:54:29.379 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:29.379 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-09T19:54:29.379 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-09T19:54:29.379 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.b 2026-03-09T19:54:29.404 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 ceph-mon[55027]: from='client.? 192.168.123.102:0/2463944290' entity='client.admin' cmd=[{"prefix": "osd getcrushmap"}]: dispatch 2026-03-09T19:54:29.607 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.b.service' 2026-03-09T19:54:29.672 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 systemd[1]: Stopping Ceph mon.b for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-b[55023]: 2026-03-09T19:54:29.488+0000 7f6a49478640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-b[55023]: 2026-03-09T19:54:29.488+0000 7f6a49478640 -1 mon.b@1(peon) e4 *** Got Signal Terminated *** 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 podman[68126]: 2026-03-09 19:54:29.524441908 +0000 UTC m=+0.050864163 container died 9f71f026ac10db56f0e2f1b2cb406c7fec8796f476cd658eb18472c432862cd5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-b, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.license=GPLv2, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, OSD_FLAVOR=default, org.label-schema.build-date=20260223) 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 podman[68126]: 2026-03-09 19:54:29.543036936 +0000 UTC m=+0.069459182 container remove 9f71f026ac10db56f0e2f1b2cb406c7fec8796f476cd658eb18472c432862cd5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-b, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, ceph=True, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, OSD_FLAVOR=default, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223) 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 bash[68126]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-mon-b 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.b.service: Deactivated successfully. 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 systemd[1]: Stopped Ceph mon.b for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:54:29.673 INFO:journalctl@ceph.mon.b.vm10.stdout:Mar 09 19:54:29 vm10 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.b.service: Consumed 1.819s CPU time. 2026-03-09T19:54:29.683 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:29.683 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-09T19:54:29.683 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-09T19:54:29.684 DEBUG:teuthology.orchestra.run.vm11:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.c 2026-03-09T19:54:29.718 DEBUG:teuthology.orchestra.run.vm11:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mon.c.service' 2026-03-09T19:54:29.803 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:29.803 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-09T19:54:29.803 INFO:tasks.cephadm.mgr.a:Stopping mgr.a... 2026-03-09T19:54:29.803 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a 2026-03-09T19:54:30.041 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.a.service' 2026-03-09T19:54:30.078 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:30.079 INFO:tasks.cephadm.mgr.a:Stopped mgr.a 2026-03-09T19:54:30.079 INFO:tasks.cephadm.mgr.b:Stopping mgr.b... 2026-03-09T19:54:30.079 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.b 2026-03-09T19:54:30.213 INFO:journalctl@ceph.mgr.b.vm10.stdout:Mar 09 19:54:30 vm10 systemd[1]: Stopping Ceph mgr.b for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:30.326 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@mgr.b.service' 2026-03-09T19:54:30.359 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:30.359 INFO:tasks.cephadm.mgr.b:Stopped mgr.b 2026-03-09T19:54:30.359 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-09T19:54:30.359 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.0 2026-03-09T19:54:30.671 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:30 vm02 systemd[1]: Stopping Ceph osd.0 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:30.671 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:30 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0[61227]: 2026-03-09T19:54:30.463+0000 7fd80a008640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:30.671 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:30 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0[61227]: 2026-03-09T19:54:30.463+0000 7fd80a008640 -1 osd.0 44 *** Got signal Terminated *** 2026-03-09T19:54:30.671 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:30 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0[61227]: 2026-03-09T19:54:30.463+0000 7fd80a008640 -1 osd.0 44 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73755]: 2026-03-09 19:54:35.503929388 +0000 UTC m=+5.054270757 container died bc19aa67d3695d7cd755b593a7eef5f9816dff4ff79e739e966736df1147a9c5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, ceph=True, org.label-schema.license=GPLv2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, io.buildah.version=1.41.3) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73755]: 2026-03-09 19:54:35.525669046 +0000 UTC m=+5.076010415 container remove bc19aa67d3695d7cd755b593a7eef5f9816dff4ff79e739e966736df1147a9c5 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 bash[73755]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73823]: 2026-03-09 19:54:35.68883225 +0000 UTC m=+0.017922426 container create 56a13c07909b4bbaf13bd464bfec55099729edb3514af839ff2fce09782f1c21 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73823]: 2026-03-09 19:54:35.73765799 +0000 UTC m=+0.066748177 container init 56a13c07909b4bbaf13bd464bfec55099729edb3514af839ff2fce09782f1c21 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0-deactivate, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, CEPH_REF=squid, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73823]: 2026-03-09 19:54:35.740985027 +0000 UTC m=+0.070075203 container start 56a13c07909b4bbaf13bd464bfec55099729edb3514af839ff2fce09782f1c21 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0-deactivate, org.label-schema.vendor=CentOS, org.label-schema.schema-version=1.0, org.label-schema.build-date=20260223, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, CEPH_REF=squid) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73823]: 2026-03-09 19:54:35.744615058 +0000 UTC m=+0.073705234 container attach 56a13c07909b4bbaf13bd464bfec55099729edb3514af839ff2fce09782f1c21 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-0-deactivate, ceph=True, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.label-schema.build-date=20260223) 2026-03-09T19:54:35.781 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 09 19:54:35 vm02 podman[73823]: 2026-03-09 19:54:35.681979117 +0000 UTC m=+0.011069293 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:54:35.928 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.0.service' 2026-03-09T19:54:35.964 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:35.964 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-09T19:54:35.964 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-09T19:54:35.964 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.1 2026-03-09T19:54:36.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:36 vm02 systemd[1]: Stopping Ceph osd.1 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:36.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:54:36.121+0000 7fdcfe733640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:36.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:54:36.121+0000 7fdcfe733640 -1 osd.1 44 *** Got signal Terminated *** 2026-03-09T19:54:36.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:36 vm02 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1[66292]: 2026-03-09T19:54:36.121+0000 7fdcfe733640 -1 osd.1 44 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[73919]: 2026-03-09 19:54:41.148599454 +0000 UTC m=+5.042207506 container died 3c7a8ac3323edd4c0e9243f9fdee99ee0d17cff255aa9069e30e1d0b5256e589 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, ceph=True, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.vendor=CentOS, CEPH_REF=squid, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0) 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[73919]: 2026-03-09 19:54:41.176894829 +0000 UTC m=+5.070502891 container remove 3c7a8ac3323edd4c0e9243f9fdee99ee0d17cff255aa9069e30e1d0b5256e589 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1, org.label-schema.license=GPLv2, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, OSD_FLAVOR=default) 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 bash[73919]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.331678924 +0000 UTC m=+0.018686706 container create f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, ceph=True, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, io.buildah.version=1.41.3, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS) 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.367027517 +0000 UTC m=+0.054035299 container init f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, org.opencontainers.image.documentation=https://docs.ceph.com/, ceph=True, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.37034804 +0000 UTC m=+0.057355822 container start f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, OSD_FLAVOR=default, ceph=True, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T19:54:41.421 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.371362098 +0000 UTC m=+0.058369880 container attach f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, ceph=True) 2026-03-09T19:54:41.662 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.1.service' 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.323009812 +0000 UTC m=+0.010017594 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.516760381 +0000 UTC m=+0.203768163 container died f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, org.label-schema.build-date=20260223, org.label-schema.schema-version=1.0, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.name=CentOS Stream 9 Base Image, OSD_FLAVOR=default, ceph=True, CEPH_REF=squid, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 podman[74000]: 2026-03-09 19:54:41.650433343 +0000 UTC m=+0.337441125 container remove f0633f773a4e1faa2dacca9994eaf9cc126fa2b8f8ae26259bbbde76119c0350 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-1-deactivate, org.label-schema.name=CentOS Stream 9 Base Image, io.buildah.version=1.41.3, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.label-schema.vendor=CentOS, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, org.label-schema.build-date=20260223) 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.1.service: Deactivated successfully. 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 systemd[1]: Stopped Ceph osd.1 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd. 2026-03-09T19:54:41.685 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 09 19:54:41 vm02 systemd[1]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.1.service: Consumed 1.237s CPU time. 2026-03-09T19:54:41.694 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:41.694 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-09T19:54:41.694 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-09T19:54:41.694 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.2 2026-03-09T19:54:42.213 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:41 vm10 systemd[1]: Stopping Ceph osd.2 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:42.213 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:41 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:54:41.801+0000 7f15b929e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:42.213 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:41 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:54:41.801+0000 7f15b929e640 -1 osd.2 44 *** Got signal Terminated *** 2026-03-09T19:54:42.213 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:41 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2[58683]: 2026-03-09T19:54:41.801+0000 7f15b929e640 -1 osd.2 44 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:46 vm10 podman[68337]: 2026-03-09 19:54:46.842270724 +0000 UTC m=+5.054014999 container died c3d22625c28a2cc7035070446c55196362f8f5df2adfd352535732bef6e841ba (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, org.opencontainers.image.documentation=https://docs.ceph.com/, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team ) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:46 vm10 podman[68337]: 2026-03-09 19:54:46.862981373 +0000 UTC m=+5.074725648 container remove c3d22625c28a2cc7035070446c55196362f8f5df2adfd352535732bef6e841ba (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, io.buildah.version=1.41.3, ceph=True, org.label-schema.build-date=20260223, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.vendor=CentOS, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:46 vm10 bash[68337]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:47 vm10 podman[68404]: 2026-03-09 19:54:47.033717134 +0000 UTC m=+0.021498774 container create db067981bce7e47b729759b83905da82e26584cf6d52011550c33f2381af4b45 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2-deactivate, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, org.opencontainers.image.authors=Ceph Release Team , FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, org.label-schema.build-date=20260223) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:47 vm10 podman[68404]: 2026-03-09 19:54:47.076815107 +0000 UTC m=+0.064596768 container init db067981bce7e47b729759b83905da82e26584cf6d52011550c33f2381af4b45 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2-deactivate, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.schema-version=1.0, CEPH_REF=squid, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, io.buildah.version=1.41.3, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.license=GPLv2, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:47 vm10 podman[68404]: 2026-03-09 19:54:47.08049139 +0000 UTC m=+0.068273041 container start db067981bce7e47b729759b83905da82e26584cf6d52011550c33f2381af4b45 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2-deactivate, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, CEPH_REF=squid, ceph=True, org.label-schema.vendor=CentOS, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, OSD_FLAVOR=default, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.build-date=20260223) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:47 vm10 podman[68404]: 2026-03-09 19:54:47.081759623 +0000 UTC m=+0.069541274 container attach db067981bce7e47b729759b83905da82e26584cf6d52011550c33f2381af4b45 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-2-deactivate, org.label-schema.vendor=CentOS, org.label-schema.license=GPLv2, org.label-schema.schema-version=1.0, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, ceph=True, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/) 2026-03-09T19:54:47.125 INFO:journalctl@ceph.osd.2.vm10.stdout:Mar 09 19:54:47 vm10 podman[68404]: 2026-03-09 19:54:47.025516823 +0000 UTC m=+0.013298484 image pull 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-09T19:54:47.248 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.2.service' 2026-03-09T19:54:47.282 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:47.282 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-09T19:54:47.282 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-09T19:54:47.282 DEBUG:teuthology.orchestra.run.vm10:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.3 2026-03-09T19:54:47.433 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:47 vm10 systemd[1]: Stopping Ceph osd.3 for 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd... 2026-03-09T19:54:47.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:47 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:54:47.432+0000 7f8a3404b640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false (PID: 1) UID: 0 2026-03-09T19:54:47.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:47 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:54:47.432+0000 7f8a3404b640 -1 osd.3 44 *** Got signal Terminated *** 2026-03-09T19:54:47.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:47 vm10 ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3[63894]: 2026-03-09T19:54:47.432+0000 7f8a3404b640 -1 osd.3 44 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68501]: 2026-03-09 19:54:52.453398581 +0000 UTC m=+5.034920736 container died c7a98648cbb4f44f4bf1b4dceec96f597bab616a1554befb95a4e08194970804 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.build-date=20260223, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.license=GPLv2, FROM_IMAGE=quay.io/centos/centos:stream9, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, OSD_FLAVOR=default, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68501]: 2026-03-09 19:54:52.481870831 +0000 UTC m=+5.063392986 container remove c7a98648cbb4f44f4bf1b4dceec96f597bab616a1554befb95a4e08194970804 (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, org.label-schema.license=GPLv2, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, OSD_FLAVOR=default, CEPH_REF=squid, ceph=True) 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 bash[68501]: ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68580]: 2026-03-09 19:54:52.640644927 +0000 UTC m=+0.018518793 container create d8fbba3b76468f9af296c9dfebf258f49893cd659f4e70c25de4544ddd70f9de (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3-deactivate, ceph=True, org.label-schema.license=GPLv2, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.label-schema.schema-version=1.0, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.vendor=CentOS, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, io.buildah.version=1.41.3, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/) 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68580]: 2026-03-09 19:54:52.674918736 +0000 UTC m=+0.052792612 container init d8fbba3b76468f9af296c9dfebf258f49893cd659f4e70c25de4544ddd70f9de (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3-deactivate, ceph=True, org.opencontainers.image.authors=Ceph Release Team , org.label-schema.name=CentOS Stream 9 Base Image, org.opencontainers.image.documentation=https://docs.ceph.com/, io.buildah.version=1.41.3, FROM_IMAGE=quay.io/centos/centos:stream9, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, CEPH_REF=squid, OSD_FLAVOR=default, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.label-schema.license=GPLv2) 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68580]: 2026-03-09 19:54:52.678572455 +0000 UTC m=+0.056446321 container start d8fbba3b76468f9af296c9dfebf258f49893cd659f4e70c25de4544ddd70f9de (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3-deactivate, org.label-schema.schema-version=1.0, org.label-schema.vendor=CentOS, io.buildah.version=1.41.3, org.label-schema.build-date=20260223, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, org.opencontainers.image.authors=Ceph Release Team , OSD_FLAVOR=default, org.label-schema.name=CentOS Stream 9 Base Image, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.license=GPLv2, CEPH_REF=squid) 2026-03-09T19:54:52.713 INFO:journalctl@ceph.osd.3.vm10.stdout:Mar 09 19:54:52 vm10 podman[68580]: 2026-03-09 19:54:52.679819138 +0000 UTC m=+0.057693004 container attach d8fbba3b76468f9af296c9dfebf258f49893cd659f4e70c25de4544ddd70f9de (image=quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc, name=ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd-osd-3-deactivate, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df, org.opencontainers.image.authors=Ceph Release Team , io.buildah.version=1.41.3, org.label-schema.schema-version=1.0, OSD_FLAVOR=default, org.opencontainers.image.documentation=https://docs.ceph.com/, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.build-date=20260223, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, ceph=True, org.label-schema.license=GPLv2, org.label-schema.vendor=CentOS, CEPH_REF=squid, org.label-schema.name=CentOS Stream 9 Base Image) 2026-03-09T19:54:52.856 DEBUG:teuthology.orchestra.run.vm10:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.3.service' 2026-03-09T19:54:52.896 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:52.896 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-09T19:54:52.896 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-09T19:54:52.896 DEBUG:teuthology.orchestra.run.vm11:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4 2026-03-09T19:54:52.936 DEBUG:teuthology.orchestra.run.vm11:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.4.service' 2026-03-09T19:54:53.015 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:53.015 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-09T19:54:53.015 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-09T19:54:53.015 DEBUG:teuthology.orchestra.run.vm11:> sudo systemctl stop ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5 2026-03-09T19:54:53.093 DEBUG:teuthology.orchestra.run.vm11:> sudo pkill -f 'journalctl -f -n 0 -u ceph-3527d82a-1bf1-11f1-92a4-57f58e2eb8dd@osd.5.service' 2026-03-09T19:54:53.169 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-09T19:54:53.169 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-09T19:54:53.169 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force --keep-logs 2026-03-09T19:54:53.308 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:54.466 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force --keep-logs 2026-03-09T19:54:54.597 INFO:teuthology.orchestra.run.vm10.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:55.730 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force --keep-logs 2026-03-09T19:54:55.857 INFO:teuthology.orchestra.run.vm11.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:56.221 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:56.249 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:56.276 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-09T19:54:56.303 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-09T19:54:56.303 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm02/crash 2026-03-09T19:54:56.303 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash -- . 2026-03-09T19:54:56.329 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash: Cannot open: No such file or directory 2026-03-09T19:54:56.329 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-09T19:54:56.330 DEBUG:teuthology.misc:Transferring archived files from vm10:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm10/crash 2026-03-09T19:54:56.330 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash -- . 2026-03-09T19:54:56.356 INFO:teuthology.orchestra.run.vm10.stderr:tar: /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash: Cannot open: No such file or directory 2026-03-09T19:54:56.356 INFO:teuthology.orchestra.run.vm10.stderr:tar: Error is not recoverable: exiting now 2026-03-09T19:54:56.358 DEBUG:teuthology.misc:Transferring archived files from vm11:/var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm11/crash 2026-03-09T19:54:56.358 DEBUG:teuthology.orchestra.run.vm11:> sudo tar c -f - -C /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash -- . 2026-03-09T19:54:56.385 INFO:teuthology.orchestra.run.vm11.stderr:tar: /var/lib/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/crash: Cannot open: No such file or directory 2026-03-09T19:54:56.385 INFO:teuthology.orchestra.run.vm11.stderr:tar: Error is not recoverable: exiting now 2026-03-09T19:54:56.386 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-09T19:54:56.386 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | egrep -v 'mons down' | egrep -v 'mon down' | egrep -v 'out of quorum' | egrep -v CEPHADM_STRAY_HOST | egrep -v CEPHADM_STRAY_DAEMON | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-09T19:54:56.424 INFO:tasks.cephadm:Compressing logs... 2026-03-09T19:54:56.424 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:54:56.467 DEBUG:teuthology.orchestra.run.vm10:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:54:56.468 DEBUG:teuthology.orchestra.run.vm11:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:54:56.490 INFO:teuthology.orchestra.run.vm10.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T19:54:56.490 INFO:teuthology.orchestra.run.vm10.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T19:54:56.491 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log 2026-03-09T19:54:56.492 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T19:54:56.492 INFO:teuthology.orchestra.run.vm02.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T19:54:56.492 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/cephadm.log: 89.9% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T19:54:56.492 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.b.log 2026-03-09T19:54:56.492 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log 2026-03-09T19:54:56.493 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.a.log 2026-03-09T19:54:56.493 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log 2026-03-09T19:54:56.494 INFO:teuthology.orchestra.run.vm11.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-09T19:54:56.494 INFO:teuthology.orchestra.run.vm11.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-09T19:54:56.495 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/cephadm.log: /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.a.log: 90.4% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T19:54:56.495 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log 2026-03-09T19:54:56.495 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.a.log 2026-03-09T19:54:56.495 INFO:teuthology.orchestra.run.vm11.stderr: 88.8% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-09T19:54:56.495 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.c.log 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log: /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.b.log: 95.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log.gz 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log: 84.7% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log.gz 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log: 82.3% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log.gz 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log 2026-03-09T19:54:56.496 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log 2026-03-09T19:54:56.497 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log: 90.5% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log.gz 2026-03-09T19:54:56.497 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.b.log 2026-03-09T19:54:56.498 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log: 84.7% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log.gz 2026-03-09T19:54:56.498 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.2.log 2026-03-09T19:54:56.499 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.c.log: 95.0% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log.gz 2026-03-09T19:54:56.499 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log 2026-03-09T19:54:56.500 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log: 80.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log.gz 2026-03-09T19:54:56.500 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log 2026-03-09T19:54:56.500 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.b.log: 90.3% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.b.log.gz 2026-03-09T19:54:56.501 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.3.log 2026-03-09T19:54:56.501 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log: 90.3% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log.gz 2026-03-09T19:54:56.501 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.4.log 2026-03-09T19:54:56.501 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log: 83.6% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.log.gz 2026-03-09T19:54:56.502 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.5.log 2026-03-09T19:54:56.502 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.a.log: gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log 2026-03-09T19:54:56.504 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log: 90.2% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.audit.log.gz 2026-03-09T19:54:56.507 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log 2026-03-09T19:54:56.508 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log: 82.5% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph.cephadm.log.gz 2026-03-09T19:54:56.513 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.0.log 2026-03-09T19:54:56.518 INFO:teuthology.orchestra.run.vm11.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.4.log: /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.5.log: 92.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.c.log.gz 2026-03-09T19:54:56.519 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log: 95.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-volume.log.gz 2026-03-09T19:54:56.519 INFO:teuthology.orchestra.run.vm11.stderr: 92.9% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.4.log.gz 2026-03-09T19:54:56.520 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.1.log 2026-03-09T19:54:56.522 INFO:teuthology.orchestra.run.vm11.stderr: 93.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.5.log.gz 2026-03-09T19:54:56.522 INFO:teuthology.orchestra.run.vm10.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.2.log: /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.3.log: 92.2% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.b.log.gz 2026-03-09T19:54:56.524 INFO:teuthology.orchestra.run.vm11.stderr: 2026-03-09T19:54:56.524 INFO:teuthology.orchestra.run.vm11.stderr:real 0m0.040s 2026-03-09T19:54:56.524 INFO:teuthology.orchestra.run.vm11.stderr:user 0m0.050s 2026-03-09T19:54:56.524 INFO:teuthology.orchestra.run.vm11.stderr:sys 0m0.019s 2026-03-09T19:54:56.535 INFO:teuthology.orchestra.run.vm10.stderr: 93.0% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.2.log.gz 2026-03-09T19:54:56.536 INFO:teuthology.orchestra.run.vm10.stderr: 93.1% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.3.log.gz 2026-03-09T19:54:56.537 INFO:teuthology.orchestra.run.vm10.stderr: 2026-03-09T19:54:56.537 INFO:teuthology.orchestra.run.vm10.stderr:real 0m0.056s 2026-03-09T19:54:56.537 INFO:teuthology.orchestra.run.vm10.stderr:user 0m0.084s 2026-03-09T19:54:56.537 INFO:teuthology.orchestra.run.vm10.stderr:sys 0m0.020s 2026-03-09T19:54:56.565 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.0.log: /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.1.log: 92.7% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.0.log.gz 2026-03-09T19:54:56.565 INFO:teuthology.orchestra.run.vm02.stderr: 89.9% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mgr.a.log.gz 2026-03-09T19:54:56.568 INFO:teuthology.orchestra.run.vm02.stderr: 93.0% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-osd.1.log.gz 2026-03-09T19:54:56.606 INFO:teuthology.orchestra.run.vm02.stderr: 91.4% -- replaced with /var/log/ceph/3527d82a-1bf1-11f1-92a4-57f58e2eb8dd/ceph-mon.a.log.gz 2026-03-09T19:54:56.607 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-09T19:54:56.608 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.127s 2026-03-09T19:54:56.608 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.178s 2026-03-09T19:54:56.608 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.027s 2026-03-09T19:54:56.608 INFO:tasks.cephadm:Archiving logs... 2026-03-09T19:54:56.608 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm02/log 2026-03-09T19:54:56.608 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T19:54:56.686 DEBUG:teuthology.misc:Transferring archived files from vm10:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm10/log 2026-03-09T19:54:56.686 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T19:54:56.716 DEBUG:teuthology.misc:Transferring archived files from vm11:/var/log/ceph to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm11/log 2026-03-09T19:54:56.716 DEBUG:teuthology.orchestra.run.vm11:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-09T19:54:56.747 INFO:tasks.cephadm:Removing cluster... 2026-03-09T19:54:56.747 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force 2026-03-09T19:54:56.884 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:57.108 DEBUG:teuthology.orchestra.run.vm10:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force 2026-03-09T19:54:57.242 INFO:teuthology.orchestra.run.vm10.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:57.482 DEBUG:teuthology.orchestra.run.vm11:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd --force 2026-03-09T19:54:57.613 INFO:teuthology.orchestra.run.vm11.stdout:Deleting cluster with fsid: 3527d82a-1bf1-11f1-92a4-57f58e2eb8dd 2026-03-09T19:54:57.824 INFO:tasks.cephadm:Removing cephadm ... 2026-03-09T19:54:57.824 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T19:54:57.842 DEBUG:teuthology.orchestra.run.vm10:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T19:54:57.858 DEBUG:teuthology.orchestra.run.vm11:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-09T19:54:57.877 INFO:tasks.cephadm:Teardown complete 2026-03-09T19:54:57.877 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-09T19:54:57.879 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-09T19:54:57.879 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T19:54:57.885 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T19:54:57.901 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-09T19:54:57.953 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T19:54:57.953 DEBUG:teuthology.orchestra.run.vm02:> 2026-03-09T19:54:57.953 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T19:54:57.953 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y remove $d || true 2026-03-09T19:54:57.953 DEBUG:teuthology.orchestra.run.vm02:> done 2026-03-09T19:54:57.960 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T19:54:57.960 DEBUG:teuthology.orchestra.run.vm10:> 2026-03-09T19:54:57.960 DEBUG:teuthology.orchestra.run.vm10:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T19:54:57.960 DEBUG:teuthology.orchestra.run.vm10:> sudo yum -y remove $d || true 2026-03-09T19:54:57.960 DEBUG:teuthology.orchestra.run.vm10:> done 2026-03-09T19:54:57.967 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-03-09T19:54:57.967 DEBUG:teuthology.orchestra.run.vm11:> 2026-03-09T19:54:57.967 DEBUG:teuthology.orchestra.run.vm11:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-03-09T19:54:57.967 DEBUG:teuthology.orchestra.run.vm11:> sudo yum -y remove $d || true 2026-03-09T19:54:57.967 DEBUG:teuthology.orchestra.run.vm11:> done 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.159 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 39 M 2026-03-09T19:54:58.160 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:54:58.162 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:54:58.162 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout:Remove 2 Packages 2026-03-09T19:54:58.169 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.170 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 39 M 2026-03-09T19:54:58.170 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:54:58.172 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:54:58.172 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:54:58.176 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:54:58.176 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:54:58.186 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:54:58.186 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout: ceph-radosgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 39 M 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout: mailcap noarch 2.1.49-5.el9 @baseos 78 k 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Remove 2 Packages 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 39 M 2026-03-09T19:54:58.191 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:54:58.194 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:54:58.194 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:54:58.209 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:54:58.209 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:54:58.210 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:54:58.219 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:54:58.242 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.243 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:58.243 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:54:58.243 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.243 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.243 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.247 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.251 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.255 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.257 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:54:58.260 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.268 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.277 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.285 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.288 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.299 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.315 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.345 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.345 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.358 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.358 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.390 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.390 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.401 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.411 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : mailcap-2.1.49-5.el9.noarch 2/2 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout: ceph-radosgw-2:19.2.3-678.ge911bdeb.el9.x86_64 mailcap-2.1.49-5.el9.noarch 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.445 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:54:58.617 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Remove 4 Packages 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 212 M 2026-03-09T19:54:58.618 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:54:58.621 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:54:58.621 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:54:58.646 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:54:58.646 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:54:58.650 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:Remove 4 Packages 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.651 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 212 M 2026-03-09T19:54:58.652 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:54:58.654 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:54:58.654 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: ceph-test x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 210 M 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: libxslt x86_64 1.1.34-12.el9 @appstream 743 k 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:54:58.662 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:58.663 INFO:teuthology.orchestra.run.vm11.stdout:Remove 4 Packages 2026-03-09T19:54:58.663 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.663 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 212 M 2026-03-09T19:54:58.663 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:54:58.665 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:54:58.665 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:54:58.679 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:54:58.680 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:54:58.691 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:54:58.691 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:54:58.710 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:54:58.715 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.717 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T19:54:58.720 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T19:54:58.736 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.746 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:54:58.752 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.754 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T19:54:58.757 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:54:58.757 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T19:54:58.763 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.765 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 2/4 2026-03-09T19:54:58.767 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libxslt-1.1.34-12.el9.x86_64 3/4 2026-03-09T19:54:58.774 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.786 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.802 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.802 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.802 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T19:54:58.802 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:58.868 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:54:58.870 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.870 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.870 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T19:54:58.870 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T19:54:58.879 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: socat-1.7.4.1-8.el9.x86_64 4/4 2026-03-09T19:54:58.879 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 1/4 2026-03-09T19:54:58.879 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libxslt-1.1.34-12.el9.x86_64 2/4 2026-03-09T19:54:58.879 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 3/4 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:58.923 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 4/4 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout: ceph-test-2:19.2.3-678.ge911bdeb.el9.x86_64 libxslt-1.1.34-12.el9.x86_64 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout: socat-1.7.4.1-8.el9.x86_64 xmlstarlet-1.6.1-20.el9.x86_64 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:58.936 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:54:59.106 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Remove 8 Packages 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 28 M 2026-03-09T19:54:59.107 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:54:59.110 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:54:59.110 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:54:59.136 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:54:59.137 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:54:59.157 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Remove 8 Packages 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 28 M 2026-03-09T19:54:59.158 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:54:59.161 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:54:59.161 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:54:59.172 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: ceph x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 0 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mds x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 7.5 M 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mon x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 18 M 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Remove 8 Packages 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 28 M 2026-03-09T19:54:59.173 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:54:59.176 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:54:59.176 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:54:59.185 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:54:59.186 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:54:59.187 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:54:59.187 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.191 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T19:54:59.193 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T19:54:59.196 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T19:54:59.199 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T19:54:59.201 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:54:59.201 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T19:54:59.201 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T19:54:59.228 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.229 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.231 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:54:59.237 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.237 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.240 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T19:54:59.242 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T19:54:59.245 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:54:59.246 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T19:54:59.248 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T19:54:59.250 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.250 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T19:54:59.254 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-03-09T19:54:59.256 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-03-09T19:54:59.259 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-03-09T19:54:59.262 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.264 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-03-09T19:54:59.265 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T19:54:59.270 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.271 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.280 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.290 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.291 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.299 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 7/8 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T19:54:59.301 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.303 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-03-09T19:54:59.319 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.320 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.372 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.376 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.376 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T19:54:59.376 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T19:54:59.376 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T19:54:59.377 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T19:54:59.377 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T19:54:59.377 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T19:54:59.403 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 8/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 1/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 3/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-03-09T19:54:59.408 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.446 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.465 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: ceph-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mds-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mon-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: lua-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: luarocks-3.9.2-5.el9.noarch 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: unzip-6.0-59.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: zip-3.0-35.el9.x86_64 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.500 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:54:59.690 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout:=========================================================================================== 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout:=========================================================================================== 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout:Removing dependent packages: 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T19:54:59.697 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T19:54:59.698 INFO:teuthology.orchestra.run.vm11.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout:=========================================================================================== 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout:Remove 102 Packages 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 613 M 2026-03-09T19:54:59.699 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:54:59.726 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:54:59.726 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:54:59.733 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:54:59.741 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T19:54:59.742 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T19:54:59.743 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout:=========================================================================================== 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout:Remove 102 Packages 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 613 M 2026-03-09T19:54:59.744 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:54:59.768 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:54:59.771 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:54:59.771 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout:=========================================================================================== 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout:=========================================================================================== 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-base x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 23 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout:Removing dependent packages: 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-immutable-object-cache x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 431 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.4 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 806 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-dashboard noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 88 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-diskprediction-local noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 66 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-rook noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 563 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-osd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 59 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-volume noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.4 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: rbd-mirror x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 85 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-grafana-dashboards noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 628 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-modules-core noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 1.5 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-prometheus-alerts noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 52 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: ceph-selinux x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 138 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: cryptsetup x86_64 2.8.1-3.el9 @baseos 770 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas x86_64 3.0.4-9.el9 @appstream 68 k 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-netlib x86_64 3.0.4-9.el9 @appstream 11 M 2026-03-09T19:54:59.774 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-openblas-openmp x86_64 3.0.4-9.el9 @appstream 39 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libcephsqlite x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 425 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libgfortran x86_64 11.5.0-14.el9 @baseos 2.8 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libquadmath x86_64 11.5.0-14.el9 @baseos 330 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libradosstriper1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.6 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: protobuf x86_64 3.14.0-17.el9 @appstream 3.5 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-compiler x86_64 3.14.0-17.el9 @crb 2.9 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-common x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 702 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-chardet noarch 4.0.0-5.el9 @anaconda 1.4 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-cheroot noarch 10.0.1-4.el9 @epel 682 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-cherrypy noarch 18.6.1-2.el9 @epel 1.1 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-cryptography x86_64 36.0.1-5.el9 @baseos 4.5 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-devel x86_64 3.9.25-3.el9 @appstream 765 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-idna noarch 2.10-7.el9.1 @anaconda 513 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jinja2 noarch 2.11.3-8.el9 @appstream 1.1 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jsonpatch noarch 1.21-16.el9 @koji-override-0 55 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-jsonpointer noarch 2.0-4.el9 @koji-override-0 34 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-logutils noarch 0.3.5-21.el9 @epel 126 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-mako noarch 1.1.4-6.el9 @appstream 534 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-03-09T19:54:59.775 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy x86_64 1:1.23.5-2.el9 @appstream 30 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9 @appstream 1.7 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-oauthlib noarch 3.1.1-5.el9 @koji-override-0 888 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pecan noarch 1.4.2-3.el9 @epel 1.3 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-ply noarch 3.11-14.el9 @baseos 430 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-prettytable noarch 0.7.2-27.el9 @koji-override-0 166 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-protobuf noarch 3.14.0-17.el9 @appstream 1.4 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1 noarch 0.4.8-7.el9 @appstream 622 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9 @appstream 1.0 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pysocks noarch 1.7.1-12.el9 @anaconda 88 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-pytz noarch 2021.1-5.el9 @koji-override-0 176 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests noarch 2.25.1-10.el9 @baseos 405 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 76 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-toml noarch 0.10.2-6.el9 @appstream 99 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-urllib3 noarch 1.26.5-7.el9 @baseos 746 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-webob noarch 1.8.8-2.el9 @epel 1.2 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-werkzeug noarch 2.0.3-3.el9.1 @epel 1.9 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: qatlib x86_64 25.08.0-2.el9 @appstream 639 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-service x86_64 25.08.0-2.el9 @appstream 69 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout:=========================================================================================== 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout:Remove 102 Packages 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 613 M 2026-03-09T19:54:59.776 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:54:59.803 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:54:59.803 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:54:59.839 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:54:59.839 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:54:59.897 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:54:59.897 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:54:59.915 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:54:59.915 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:00.018 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:00.018 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.025 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.045 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:00.046 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.046 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.054 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.062 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.069 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:00.069 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.075 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.076 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.077 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 1/102 2026-03-09T19:55:00.090 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.090 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T19:55:00.091 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-03-09T19:55:00.097 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.098 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.114 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:00.114 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T19:55:00.114 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.138 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 3/102 2026-03-09T19:55:00.138 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.151 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.159 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T19:55:00.164 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T19:55:00.164 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.173 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.175 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.182 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T19:55:00.182 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T19:55:00.186 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T19:55:00.186 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T19:55:00.186 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.194 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T19:55:00.197 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T19:55:00.200 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 4/102 2026-03-09T19:55:00.201 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.209 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T19:55:00.210 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 5/102 2026-03-09T19:55:00.214 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T19:55:00.215 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 6/102 2026-03-09T19:55:00.215 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T19:55:00.221 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.223 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T19:55:00.227 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T19:55:00.227 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.227 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:00.235 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cherrypy-18.6.1-2.el9.noarch 8/102 2026-03-09T19:55:00.236 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.239 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cheroot-10.0.1-4.el9.noarch 9/102 2026-03-09T19:55:00.248 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 10/102 2026-03-09T19:55:00.252 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 11/102 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T19:55:00.253 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.254 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.254 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.254 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:55:00.254 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.259 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.262 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.269 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.275 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-03-09T19:55:00.278 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.284 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.284 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T19:55:00.289 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T19:55:00.289 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.289 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.289 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:55:00.289 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.294 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:00.298 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T19:55:00.298 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.310 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.311 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T19:55:00.313 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.313 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.313 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-03-09T19:55:00.313 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.313 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T19:55:00.318 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T19:55:00.318 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T19:55:00.322 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.323 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T19:55:00.329 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T19:55:00.331 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T19:55:00.333 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 13/102 2026-03-09T19:55:00.335 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 14/102 2026-03-09T19:55:00.337 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T19:55:00.344 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 15/102 2026-03-09T19:55:00.345 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T19:55:00.348 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jinja2-2.11.3-8.el9.noarch 16/102 2026-03-09T19:55:00.352 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T19:55:00.358 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-requests-2.25.1-10.el9.noarch 17/102 2026-03-09T19:55:00.361 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T19:55:00.368 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T19:55:00.370 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T19:55:00.372 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 18/102 2026-03-09T19:55:00.378 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T19:55:00.379 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pecan-1.4.2-3.el9.noarch 19/102 2026-03-09T19:55:00.382 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T19:55:00.389 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 20/102 2026-03-09T19:55:00.391 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T19:55:00.396 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9.noarch 21/102 2026-03-09T19:55:00.398 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T19:55:00.401 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T19:55:00.402 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.405 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T19:55:00.408 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T19:55:00.409 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.418 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T19:55:00.427 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-urllib3-1.26.5-7.el9.noarch 22/102 2026-03-09T19:55:00.429 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T19:55:00.429 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.434 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 23/102 2026-03-09T19:55:00.437 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 24/102 2026-03-09T19:55:00.439 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.446 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 25/102 2026-03-09T19:55:00.456 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 26/102 2026-03-09T19:55:00.456 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.464 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 27/102 2026-03-09T19:55:00.508 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T19:55:00.521 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T19:55:00.535 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.535 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T19:55:00.535 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.536 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.539 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T19:55:00.555 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T19:55:00.567 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.571 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.571 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T19:55:00.571 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.572 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.573 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 28/102 2026-03-09T19:55:00.583 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T19:55:00.590 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 29/102 2026-03-09T19:55:00.590 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T19:55:00.593 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T19:55:00.595 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T19:55:00.604 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.608 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.608 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-03-09T19:55:00.608 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.609 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.619 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.620 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.622 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T19:55:00.629 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T19:55:00.633 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T19:55:00.636 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.637 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 30/102 2026-03-09T19:55:00.637 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T19:55:00.640 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T19:55:00.642 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T19:55:00.645 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T19:55:00.648 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T19:55:00.651 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T19:55:00.653 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 31/102 2026-03-09T19:55:00.656 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T19:55:00.658 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cryptography-36.0.1-5.el9.x86_64 32/102 2026-03-09T19:55:00.660 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : protobuf-compiler-3.14.0-17.el9.x86_64 33/102 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.661 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.663 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.663 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 34/102 2026-03-09T19:55:00.675 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.679 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T19:55:00.682 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-03-09T19:55:00.684 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.685 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T19:55:00.685 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.688 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T19:55:00.691 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T19:55:00.696 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T19:55:00.696 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 35/102 2026-03-09T19:55:00.700 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-mako-1.1.4-6.el9.noarch 36/102 2026-03-09T19:55:00.701 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T19:55:00.702 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 37/102 2026-03-09T19:55:00.705 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 38/102 2026-03-09T19:55:00.707 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 39/102 2026-03-09T19:55:00.709 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T19:55:00.711 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 40/102 2026-03-09T19:55:00.747 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 41/102 2026-03-09T19:55:00.749 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T19:55:00.749 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T19:55:00.751 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 42/102 2026-03-09T19:55:00.751 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T19:55:00.756 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T19:55:00.758 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T19:55:00.762 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T19:55:00.763 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T19:55:00.764 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T19:55:00.766 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T19:55:00.771 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T19:55:00.773 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T19:55:00.777 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T19:55:00.780 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T19:55:00.785 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.785 INFO:teuthology.orchestra.run.vm11.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.785 INFO:teuthology.orchestra.run.vm11.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:55:00.785 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.785 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.792 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.793 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T19:55:00.795 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T19:55:00.798 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T19:55:00.799 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.799 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.799 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:55:00.799 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.800 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 43/102 2026-03-09T19:55:00.800 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T19:55:00.800 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.802 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T19:55:00.804 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T19:55:00.807 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T19:55:00.809 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.812 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T19:55:00.812 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T19:55:00.813 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-numpy-1:1.23.5-2.el9.x86_64 44/102 2026-03-09T19:55:00.815 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T19:55:00.815 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : flexiblas-netlib-3.0.4-9.el9.x86_64 45/102 2026-03-09T19:55:00.819 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T19:55:00.819 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T19:55:00.821 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 46/102 2026-03-09T19:55:00.821 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T19:55:00.823 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T19:55:00.824 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 47/102 2026-03-09T19:55:00.824 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T19:55:00.825 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T19:55:00.827 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T19:55:00.828 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libgfortran-11.5.0-14.el9.x86_64 48/102 2026-03-09T19:55:00.828 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T19:55:00.830 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T19:55:00.832 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T19:55:00.832 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 49/102 2026-03-09T19:55:00.833 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T19:55:00.837 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T19:55:00.841 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T19:55:00.841 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T19:55:00.845 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T19:55:00.846 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T19:55:00.847 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T19:55:00.849 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T19:55:00.850 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T19:55:00.852 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T19:55:00.854 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.854 INFO:teuthology.orchestra.run.vm10.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-03-09T19:55:00.854 INFO:teuthology.orchestra.run.vm10.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-03-09T19:55:00.854 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:00.855 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.855 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T19:55:00.857 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T19:55:00.858 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T19:55:00.865 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-immutable-object-cache-2:19.2.3-678.ge911bd 50/102 2026-03-09T19:55:00.865 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T19:55:00.865 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T19:55:00.867 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 51/102 2026-03-09T19:55:00.868 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T19:55:00.870 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T19:55:00.870 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : flexiblas-3.0.4-9.el9.x86_64 52/102 2026-03-09T19:55:00.874 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-ply-3.11-14.el9.noarch 53/102 2026-03-09T19:55:00.874 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T19:55:00.874 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T19:55:00.876 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 54/102 2026-03-09T19:55:00.877 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T19:55:00.879 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 55/102 2026-03-09T19:55:00.880 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T19:55:00.881 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T19:55:00.882 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 56/102 2026-03-09T19:55:00.883 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T19:55:00.886 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-toml-0.10.2-6.el9.noarch 57/102 2026-03-09T19:55:00.886 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T19:55:00.889 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T19:55:00.890 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 58/102 2026-03-09T19:55:00.891 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T19:55:00.895 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T19:55:00.897 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T19:55:00.898 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 59/102 2026-03-09T19:55:00.899 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T19:55:00.901 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T19:55:00.902 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T19:55:00.903 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T19:55:00.903 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-devel-3.9.25-3.el9.x86_64 60/102 2026-03-09T19:55:00.905 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T19:55:00.906 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-jsonpointer-2.0-4.el9.noarch 61/102 2026-03-09T19:55:00.909 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 62/102 2026-03-09T19:55:00.911 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T19:55:00.912 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-idna-2.10-7.el9.1.noarch 63/102 2026-03-09T19:55:00.912 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T19:55:00.919 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T19:55:00.921 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T19:55:00.923 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pysocks-1.7.1-12.el9.noarch 64/102 2026-03-09T19:55:00.925 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T19:55:00.928 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T19:55:00.928 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-pyasn1-0.4.8-7.el9.noarch 65/102 2026-03-09T19:55:00.929 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T19:55:00.933 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-logutils-0.3.5-21.el9.noarch 66/102 2026-03-09T19:55:00.935 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T19:55:00.938 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:00.938 INFO:teuthology.orchestra.run.vm11.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T19:55:00.938 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:00.945 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:00.952 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T19:55:00.953 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-webob-1.8.8-2.el9.noarch 67/102 2026-03-09T19:55:00.960 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 68/102 2026-03-09T19:55:00.964 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 69/102 2026-03-09T19:55:00.968 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 70/102 2026-03-09T19:55:00.971 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 71/102 2026-03-09T19:55:00.973 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:00.973 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T19:55:00.973 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:00.978 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 72/102 2026-03-09T19:55:00.978 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:00.978 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:00.981 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:00.983 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-protobuf-3.14.0-17.el9.noarch 73/102 2026-03-09T19:55:00.986 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 74/102 2026-03-09T19:55:00.991 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:00.995 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 75/102 2026-03-09T19:55:00.996 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T19:55:01.000 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T19:55:01.001 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 76/102 2026-03-09T19:55:01.002 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T19:55:01.002 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:01.005 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 77/102 2026-03-09T19:55:01.008 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 78/102 2026-03-09T19:55:01.009 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:01.010 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:01.010 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 79/102 2026-03-09T19:55:01.016 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 80/102 2026-03-09T19:55:01.020 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-werkzeug-2.0.3-3.el9.1.noarch 81/102 2026-03-09T19:55:01.021 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:01.027 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T19:55:01.030 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T19:55:01.032 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T19:55:01.032 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:01.041 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:01.041 INFO:teuthology.orchestra.run.vm10.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-03-09T19:55:01.041 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:01.049 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:01.079 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 82/102 2026-03-09T19:55:01.079 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:01.091 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 83/102 2026-03-09T19:55:01.097 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 84/102 2026-03-09T19:55:01.100 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 85/102 2026-03-09T19:55:01.103 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-prettytable-0.7.2-27.el9.noarch 86/102 2026-03-09T19:55:01.103 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:06.680 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-03-09T19:55:06.681 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:06.689 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T19:55:06.709 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.709 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.718 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.720 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T19:55:06.723 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T19:55:06.725 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T19:55:06.727 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T19:55:06.727 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.743 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.745 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T19:55:06.747 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T19:55:06.750 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T19:55:06.753 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T19:55:06.758 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T19:55:06.765 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T19:55:06.770 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T19:55:06.770 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T19:55:06.871 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T19:55:06.872 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T19:55:06.874 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T19:55:06.875 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /sys 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /proc 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /mnt 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /var/tmp 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /home 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /root 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout:skipping the directory /tmp 2026-03-09T19:55:06.903 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 87/102 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /sys 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /proc 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /mnt 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /var/tmp 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /home 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /root 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout:skipping the directory /tmp 2026-03-09T19:55:06.912 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:06.913 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T19:55:06.920 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : qatlib-25.08.0-2.el9.x86_64 88/102 2026-03-09T19:55:06.931 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.931 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.939 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.939 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.940 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.944 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T19:55:06.948 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T19:55:06.949 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: qatlib-service-25.08.0-2.el9.x86_64 89/102 2026-03-09T19:55:06.952 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T19:55:06.952 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : gperftools-libs-2.9.1-3.el9.x86_64 90/102 2026-03-09T19:55:06.955 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T19:55:06.955 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libunwind-1.6.2-1.el9.x86_64 91/102 2026-03-09T19:55:06.955 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.957 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : pciutils-3.7.0-7.el9.x86_64 92/102 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : liboath-2.6.12-1.el9.x86_64 93/102 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:55:06.960 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:55:06.961 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:06.962 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:06.973 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.976 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 94/102 2026-03-09T19:55:06.976 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T19:55:06.979 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ledmon-libs-1.1.0-3.el9.x86_64 95/102 2026-03-09T19:55:06.980 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T19:55:06.981 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libquadmath-11.5.0-14.el9.x86_64 96/102 2026-03-09T19:55:06.984 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T19:55:06.985 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-markupsafe-1.1.1-12.el9.x86_64 97/102 2026-03-09T19:55:06.987 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T19:55:06.988 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : protobuf-3.14.0-17.el9.x86_64 98/102 2026-03-09T19:55:06.994 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T19:55:06.996 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libconfig-1.7.2-9.el9.x86_64 99/102 2026-03-09T19:55:07.003 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T19:55:07.004 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : cryptsetup-2.8.1-3.el9.x86_64 100/102 2026-03-09T19:55:07.008 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T19:55:07.008 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.009 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : abseil-cpp-20211102.0-4.el9.x86_64 101/102 2026-03-09T19:55:07.009 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:07.118 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T19:55:07.120 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T19:55:07.121 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T19:55:07.122 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 1/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 3/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.e 4/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-immutable-object-cache-2:19.2.3-678.ge911bd 5/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 6/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noar 7/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.no 8/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-diskprediction-local-2:19.2.3-678.ge911 9/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9 10/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 11/102 2026-03-09T19:55:07.132 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 12/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el 13/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 14/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 15/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : cryptsetup-2.8.1-3.el9.x86_64 16/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-3.0.4-9.el9.x86_64 17/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-netlib-3.0.4-9.el9.x86_64 18/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 19/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 20/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 21/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 22/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 23/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 24/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libgfortran-11.5.0-14.el9.x86_64 25/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 26/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libquadmath-11.5.0-14.el9.x86_64 27/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_ 28/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 29/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 30/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 31/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 32/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 33/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : protobuf-3.14.0-17.el9.x86_64 34/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : protobuf-compiler-3.14.0-17.el9.x86_64 35/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 36/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 37/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 38/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 39/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 40/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 41/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x 42/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 43/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 44/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-chardet-4.0.0-5.el9.noarch 45/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cheroot-10.0.1-4.el9.noarch 46/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cherrypy-18.6.1-2.el9.noarch 47/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cryptography-36.0.1-5.el9.x86_64 48/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-devel-3.9.25-3.el9.x86_64 49/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 50/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 51/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 52/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-idna-2.10-7.el9.1.noarch 53/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 54/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 55/102 2026-03-09T19:55:07.135 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 56/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 57/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 58/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 59/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jinja2-2.11.3-8.el9.noarch 60/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jsonpatch-1.21-16.el9.noarch 61/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-jsonpointer-2.0-4.el9.noarch 62/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 63/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 64/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-logutils-0.3.5-21.el9.noarch 65/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-mako-1.1.4-6.el9.noarch 66/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 67/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 68/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 69/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-numpy-1:1.23.5-2.el9.x86_64 70/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9.x86_64 71/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-oauthlib-3.1.1-5.el9.noarch 72/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 73/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pecan-1.4.2-3.el9.noarch 74/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ply-3.11-14.el9.noarch 75/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 76/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-prettytable-0.7.2-27.el9.noarch 77/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-protobuf-3.14.0-17.el9.noarch 78/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 79/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyasn1-0.4.8-7.el9.noarch 80/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9.noarch 81/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 82/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pysocks-1.7.1-12.el9.noarch 83/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-pytz-2021.1-5.el9.noarch 84/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 85/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-requests-2.25.1-10.el9.noarch 86/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 87/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 88/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 89/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 90/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 91/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-toml-0.10.2-6.el9.noarch 92/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 93/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 94/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-webob-1.8.8-2.el9.noarch 95/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 96/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-werkzeug-2.0.3-3.el9.1.noarch 97/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 98/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatlib-25.08.0-2.el9.x86_64 99/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatlib-service-25.08.0-2.el9.x86_64 100/102 2026-03-09T19:55:07.136 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 101/102 2026-03-09T19:55:07.172 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:Remove 1 Package 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 775 k 2026-03-09T19:55:07.173 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:55:07.175 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:55:07.175 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:55:07.176 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:55:07.176 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:55:07.192 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:07.192 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.206 INFO:teuthology.orchestra.run.vm11.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:55:07.207 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:55:07.208 INFO:teuthology.orchestra.run.vm11.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.209 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:07.217 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 102/102 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-base-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-grafana-dashboards-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-immutable-object-cache-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-dashboard-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-diskprediction-local-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-modules-core-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-mgr-rook-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-osd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-prometheus-alerts-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-selinux-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ceph-volume-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: cryptsetup-2.8.1-3.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-netlib-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: flexiblas-openblas-openmp-3.0.4-9.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: grpc-data-1.46.7-10.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libcephsqlite-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libconfig-1.7.2-9.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libgfortran-11.5.0-14.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: liboath-2.6.12-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libquadmath-11.5.0-14.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libradosstriper1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: libunwind-1.6.2-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: openblas-0.3.29-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: pciutils-3.7.0-7.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-3.14.0-17.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: protobuf-compiler-3.14.0-17.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-babel-2.9.1-2.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-common-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-chardet-4.0.0-5.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-cheroot-10.0.1-4.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-cherrypy-18.6.1-2.el9.noarch 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-cryptography-36.0.1-5.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-devel-3.9.25-3.el9.x86_64 2026-03-09T19:55:07.218 INFO:teuthology.orchestra.run.vm10.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-idna-2.10-7.el9.1.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jinja2-2.11.3-8.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jsonpatch-1.21-16.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-jsonpointer-2.0-4.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-logutils-0.3.5-21.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-mako-1.1.4-6.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-numpy-f2py-1:1.23.5-2.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-oauthlib-3.1.1-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-packaging-20.9-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pecan-1.4.2-3.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-ply-3.11-14.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-portend-3.1.0-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-prettytable-0.7.2-27.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-protobuf-3.14.0-17.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-0.4.8-7.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pyasn1-modules-0.4.8-7.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pycparser-2.20-6.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pysocks-1.7.1-12.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-pytz-2021.1-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-2.25.1-10.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-routes-2.5.1-5.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-rsa-4.9-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-toml-0.10.2-6.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-webob-1.8.8-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-werkzeug-2.0.3-3.el9.1.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-25.08.0-2.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: qatlib-service-25.08.0-2.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-03-09T19:55:07.219 INFO:teuthology.orchestra.run.vm10.stdout: rbd-mirror-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:07.220 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.220 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:07.307 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:07.351 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:07.414 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:Remove 1 Package 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 775 k 2026-03-09T19:55:07.415 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:55:07.417 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:55:07.417 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:55:07.418 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:55:07.418 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:55:07.434 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:07.435 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout: cephadm noarch 2:19.2.3-678.ge911bdeb.el9 @ceph-noarch 775 k 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Remove 1 Package 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 775 k 2026-03-09T19:55:07.443 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:55:07.445 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:55:07.445 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:55:07.446 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:55:07.446 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:07.463 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:07.463 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.556 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T19:55:07.556 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:07.559 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:07.560 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:07.560 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:07.569 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.584 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:07.611 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 1/1 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout: cephadm-2:19.2.3-678.ge911bdeb.el9.noarch 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:07.626 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:07.733 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr 2026-03-09T19:55:07.733 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:07.736 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:07.737 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:07.737 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:07.807 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T19:55:07.807 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:07.811 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:07.811 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:07.811 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:07.817 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-immutable-object-cache 2026-03-09T19:55:07.817 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:07.821 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:07.821 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:07.821 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:07.919 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T19:55:07.919 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:07.922 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:07.923 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:07.923 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:08.010 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-mgr 2026-03-09T19:55:08.010 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:08.014 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.014 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:08.014 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:08.029 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-mgr 2026-03-09T19:55:08.029 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:08.032 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.033 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:08.033 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:08.111 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T19:55:08.112 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:08.115 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:08.115 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:08.115 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:08.199 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T19:55:08.199 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:08.203 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.204 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:08.204 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:08.216 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-mgr-dashboard 2026-03-09T19:55:08.217 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:08.222 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.223 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:08.223 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:08.296 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-rook 2026-03-09T19:55:08.296 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:08.299 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:08.299 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:08.299 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:08.391 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T19:55:08.391 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:08.395 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.396 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:08.396 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:08.407 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-mgr-diskprediction-local 2026-03-09T19:55:08.408 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:08.411 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.411 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:08.411 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:08.467 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T19:55:08.467 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:08.470 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:08.470 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:08.470 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:08.571 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-mgr-rook 2026-03-09T19:55:08.571 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:08.575 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.576 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:08.576 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:08.586 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-mgr-rook 2026-03-09T19:55:08.586 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:08.590 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.590 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:08.590 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Remove 1 Package 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.6 M 2026-03-09T19:55:08.653 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:55:08.655 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:55:08.655 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:55:08.665 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:55:08.665 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:55:08.691 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:08.705 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:08.761 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T19:55:08.761 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:08.765 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.765 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:08.766 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:08.773 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-mgr-cephadm 2026-03-09T19:55:08.773 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:08.777 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.777 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:08.777 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:08.782 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:08.830 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:08.830 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:08.830 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:08.831 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:08.831 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:08.831 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:08.963 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:Remove 1 Package 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 3.6 M 2026-03-09T19:55:08.964 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:55:08.966 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:55:08.966 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:55:08.974 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout: ceph-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.6 M 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:Remove 1 Package 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 3.6 M 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:55:08.975 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:55:08.976 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:55:08.976 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:55:08.976 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:55:08.987 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:55:08.987 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:09.002 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:09.016 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.018 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:09.023 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: ceph-volume 2026-03-09T19:55:09.023 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:09.027 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:09.027 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:09.027 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:09.033 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.097 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.100 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.146 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 1/1 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout: ceph-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.151 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:09.237 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 610 k 2026-03-09T19:55:09.238 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:55:09.240 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:55:09.240 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:55:09.251 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:55:09.251 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:55:09.283 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:09.285 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.299 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.339 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: ceph-volume 2026-03-09T19:55:09.339 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:09.342 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:09.343 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:09.343 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:09.360 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: ceph-volume 2026-03-09T19:55:09.360 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:09.363 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:09.364 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:09.364 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:09.370 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.370 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.424 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout:Removing dependent packages: 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T19:55:09.527 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout:Remove 2 Packages 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 610 k 2026-03-09T19:55:09.528 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:55:09.529 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:55:09.529 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:55:09.541 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:55:09.541 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout: librados-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 456 k 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:Removing dependent packages: 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs-devel x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 153 k 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.549 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:55:09.550 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.550 INFO:teuthology.orchestra.run.vm10.stdout:Remove 2 Packages 2026-03-09T19:55:09.550 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.550 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 610 k 2026-03-09T19:55:09.550 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:55:09.551 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:55:09.551 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:55:09.562 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:55:09.562 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:09.569 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:09.572 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.586 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.589 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:09.591 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.605 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.629 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:09.629 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Remove 3 Packages 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 3.7 M 2026-03-09T19:55:09.630 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:55:09.633 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:55:09.633 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:55:09.653 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:55:09.654 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:55:09.667 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.667 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.687 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:09.689 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:09.689 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.689 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 1/2 2026-03-09T19:55:09.691 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:09.691 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.715 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2/2 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout: librados-devel-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.733 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:09.768 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:09.768 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:09.768 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:09.807 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Removing dependent packages: 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Remove 3 Packages 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 3.7 M 2026-03-09T19:55:09.925 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:55:09.927 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:55:09.927 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:55:09.946 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:55:09.946 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:55:09.951 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repo Size 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 3.0 M 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Removing dependent packages: 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: python3-cephfs x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 514 k 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-argparse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 187 k 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Remove 3 Packages 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 3.7 M 2026-03-09T19:55:09.952 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:55:09.954 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:55:09.954 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:55:09.974 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:55:09.974 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:09.980 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:09.983 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:09.985 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:09.985 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:09.999 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: libcephfs-devel 2026-03-09T19:55:09.999 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:10.003 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:10.003 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:10.004 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:10.007 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:10.010 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:10.012 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:10.012 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:10.054 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:10.055 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:10.055 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:10.079 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:10.079 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 1/3 2026-03-09T19:55:10.079 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86 2/3 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.094 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 3/3 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: libcephfs2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: python3-ceph-argparse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: python3-cephfs-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.123 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:10.193 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T19:55:10.194 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:Remove 20 Packages 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 79 M 2026-03-09T19:55:10.195 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-03-09T19:55:10.199 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-03-09T19:55:10.199 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-03-09T19:55:10.221 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-03-09T19:55:10.222 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-03-09T19:55:10.265 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-03-09T19:55:10.268 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T19:55:10.270 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T19:55:10.273 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T19:55:10.274 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.285 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: libcephfs-devel 2026-03-09T19:55:10.285 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:10.288 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.289 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:10.289 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:10.289 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:10.290 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T19:55:10.292 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T19:55:10.294 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.296 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T19:55:10.299 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T19:55:10.299 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.311 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: libcephfs-devel 2026-03-09T19:55:10.311 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:10.314 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.314 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.315 INFO:teuthology.orchestra.run.vm02.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T19:55:10.315 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:10.315 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:10.316 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:10.316 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:10.331 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.334 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.337 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T19:55:10.340 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T19:55:10.343 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T19:55:10.345 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T19:55:10.347 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T19:55:10.349 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T19:55:10.350 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T19:55:10.364 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.431 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T19:55:10.432 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-09T19:55:10.474 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:10.496 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout: Package Arch Version Repository Size 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout:Removing: 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout:Removing dependent packages: 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.497 INFO:teuthology.orchestra.run.vm11.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:Removing unused dependencies: 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:Transaction Summary 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:================================================================================ 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:Remove 20 Packages 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:Freed space: 79 M 2026-03-09T19:55:10.498 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction check 2026-03-09T19:55:10.502 INFO:teuthology.orchestra.run.vm11.stdout:Transaction check succeeded. 2026-03-09T19:55:10.502 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction test 2026-03-09T19:55:10.517 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: Package Arch Version Repository Size 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Removing: 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: librados2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 12 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Removing dependent packages: 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: python3-rados x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: python3-rbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 1.1 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: python3-rgw x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 265 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: qemu-kvm-block-rbd x86_64 17:10.1.0-15.el9 @appstream 37 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: rbd-fuse x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 227 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: rbd-nbd x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 490 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Removing unused dependencies: 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: boost-program-options x86_64 1.75.0-13.el9 @appstream 276 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: libarrow x86_64 9.0.0-15.el9 @epel 18 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-doc noarch 9.0.0-15.el9 @epel 122 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: libnbd x86_64 1.20.3-4.el9 @appstream 453 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: libpmemobj x86_64 1.12.1-1.el9 @appstream 383 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: librabbitmq x86_64 0.11.0-7.el9 @appstream 102 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: librbd1 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 13 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: librdkafka x86_64 1.6.1-102.el9 @appstream 2.0 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: librgw2 x86_64 2:19.2.3-678.ge911bdeb.el9 @ceph 19 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: lttng-ust x86_64 2.12.0-6.el9 @appstream 1.0 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: parquet-libs x86_64 9.0.0-15.el9 @epel 2.8 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: re2 x86_64 1:20211101-20.el9 @epel 472 k 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: thrift x86_64 0.15.0-4.el9 @epel 4.8 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Transaction Summary 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:================================================================================ 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Remove 20 Packages 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Freed space: 79 M 2026-03-09T19:55:10.519 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction check 2026-03-09T19:55:10.524 INFO:teuthology.orchestra.run.vm10.stdout:Transaction check succeeded. 2026-03-09T19:55:10.524 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction test 2026-03-09T19:55:10.525 INFO:teuthology.orchestra.run.vm11.stdout:Transaction test succeeded. 2026-03-09T19:55:10.526 INFO:teuthology.orchestra.run.vm11.stdout:Running transaction 2026-03-09T19:55:10.547 INFO:teuthology.orchestra.run.vm10.stdout:Transaction test succeeded. 2026-03-09T19:55:10.548 INFO:teuthology.orchestra.run.vm10.stdout:Running transaction 2026-03-09T19:55:10.570 INFO:teuthology.orchestra.run.vm11.stdout: Preparing : 1/1 2026-03-09T19:55:10.573 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T19:55:10.575 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T19:55:10.579 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T19:55:10.579 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.591 INFO:teuthology.orchestra.run.vm10.stdout: Preparing : 1/1 2026-03-09T19:55:10.593 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.594 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 1/20 2026-03-09T19:55:10.596 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T19:55:10.597 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2/20 2026-03-09T19:55:10.598 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T19:55:10.600 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.601 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 3/20 2026-03-09T19:55:10.601 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.603 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T19:55:10.606 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T19:55:10.606 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.615 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 4/20 2026-03-09T19:55:10.618 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : parquet-libs-9.0.0-15.el9.x86_64 5/20 2026-03-09T19:55:10.620 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 6/20 2026-03-09T19:55:10.622 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.622 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.622 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.622 INFO:teuthology.orchestra.run.vm11.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T19:55:10.622 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.624 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 8/20 2026-03-09T19:55:10.627 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libarrow-doc-9.0.0-15.el9.noarch 9/20 2026-03-09T19:55:10.627 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.639 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.641 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.643 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.643 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.643 INFO:teuthology.orchestra.run.vm10.stdout:warning: file /etc/ceph: remove failed: No such file or directory 2026-03-09T19:55:10.643 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.644 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T19:55:10.648 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T19:55:10.650 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T19:55:10.653 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T19:55:10.654 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T19:55:10.656 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T19:55:10.658 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T19:55:10.660 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 11/20 2026-03-09T19:55:10.662 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libarrow-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.666 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : re2-1:20211101-20.el9.x86_64 13/20 2026-03-09T19:55:10.670 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : lttng-ust-2.12.0-6.el9.x86_64 14/20 2026-03-09T19:55:10.672 INFO:teuthology.orchestra.run.vm11.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.673 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : thrift-0.15.0-4.el9.x86_64 15/20 2026-03-09T19:55:10.676 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libnbd-1.20.3-4.el9.x86_64 16/20 2026-03-09T19:55:10.677 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : libpmemobj-1.12.1-1.el9.x86_64 17/20 2026-03-09T19:55:10.679 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : boost-program-options-1.75.0-13.el9.x86_64 18/20 2026-03-09T19:55:10.681 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librabbitmq-0.11.0-7.el9.x86_64 19/20 2026-03-09T19:55:10.693 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: librbd1 2026-03-09T19:55:10.693 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:10.695 INFO:teuthology.orchestra.run.vm10.stdout: Erasing : librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.696 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:10.696 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:10.696 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.749 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T19:55:10.750 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Running scriptlet: librdkafka-1.6.1-102.el9.x86_64 20/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : boost-program-options-1.75.0-13.el9.x86_64 1/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 2/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 3/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 4/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 5/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 6/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 7/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 8/20 2026-03-09T19:55:10.763 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 9/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 10/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 11/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 12/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 13/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 14/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 15/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 16/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 17/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 18/20 2026-03-09T19:55:10.764 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : re2-1:20211101-20.el9.x86_64 19/20 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout:Removed: 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout: 2026-03-09T19:55:10.802 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 20/20 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout:Removed: 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: boost-program-options-1.75.0-13.el9.x86_64 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: libnbd-1.20.3-4.el9.x86_64 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-03-09T19:55:10.813 INFO:teuthology.orchestra.run.vm10.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: librados2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: librbd1-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: librgw2-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: python3-rados-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: python3-rbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: python3-rgw-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: qemu-kvm-block-rbd-17:10.1.0-15.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: rbd-fuse-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: rbd-nbd-2:19.2.3-678.ge911bdeb.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: re2-1:20211101-20.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: thrift-0.15.0-4.el9.x86_64 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout: 2026-03-09T19:55:10.814 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:10.873 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rados 2026-03-09T19:55:10.873 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:10.875 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:10.876 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:10.876 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.034 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: librbd1 2026-03-09T19:55:11.035 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.035 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: librbd1 2026-03-09T19:55:11.035 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.038 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.038 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.038 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.039 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.039 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.039 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.063 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rgw 2026-03-09T19:55:11.063 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.065 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.066 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.066 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.234 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: python3-rados 2026-03-09T19:55:11.235 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.237 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.238 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.238 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.249 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: python3-rados 2026-03-09T19:55:11.249 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.251 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-cephfs 2026-03-09T19:55:11.251 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.251 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.252 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.252 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.254 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.254 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.254 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.425 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: python3-rgw 2026-03-09T19:55:11.425 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.426 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: python3-rgw 2026-03-09T19:55:11.426 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.427 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.427 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.428 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.428 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.428 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.428 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.444 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: python3-rbd 2026-03-09T19:55:11.444 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.446 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.447 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.447 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.593 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: python3-cephfs 2026-03-09T19:55:11.593 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.595 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.595 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.595 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.599 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: python3-cephfs 2026-03-09T19:55:11.600 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.602 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.603 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.603 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.622 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-fuse 2026-03-09T19:55:11.622 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.624 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.625 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.625 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.755 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: python3-rbd 2026-03-09T19:55:11.755 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.756 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.757 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.757 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.778 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: python3-rbd 2026-03-09T19:55:11.778 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.780 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.780 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.780 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.789 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-mirror 2026-03-09T19:55:11.789 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.791 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.792 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.792 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.925 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: rbd-fuse 2026-03-09T19:55:11.925 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:11.927 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:11.928 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:11.928 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:11.947 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: rbd-fuse 2026-03-09T19:55:11.947 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:11.950 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:11.951 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:11.951 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:11.964 INFO:teuthology.orchestra.run.vm02.stdout:No match for argument: rbd-nbd 2026-03-09T19:55:11.964 INFO:teuthology.orchestra.run.vm02.stderr:No packages marked for removal. 2026-03-09T19:55:11.966 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-03-09T19:55:11.967 INFO:teuthology.orchestra.run.vm02.stdout:Nothing to do. 2026-03-09T19:55:11.967 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-03-09T19:55:11.995 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-03-09T19:55:12.114 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: rbd-mirror 2026-03-09T19:55:12.115 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:12.117 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:12.117 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:12.117 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:12.124 INFO:teuthology.orchestra.run.vm02.stdout:56 files removed 2026-03-09T19:55:12.133 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: rbd-mirror 2026-03-09T19:55:12.133 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:12.136 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:12.137 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:12.137 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:12.149 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.178 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean expire-cache 2026-03-09T19:55:12.300 INFO:teuthology.orchestra.run.vm10.stdout:No match for argument: rbd-nbd 2026-03-09T19:55:12.300 INFO:teuthology.orchestra.run.vm10.stderr:No packages marked for removal. 2026-03-09T19:55:12.302 INFO:teuthology.orchestra.run.vm10.stdout:Dependencies resolved. 2026-03-09T19:55:12.302 INFO:teuthology.orchestra.run.vm10.stdout:Nothing to do. 2026-03-09T19:55:12.303 INFO:teuthology.orchestra.run.vm10.stdout:Complete! 2026-03-09T19:55:12.320 INFO:teuthology.orchestra.run.vm11.stdout:No match for argument: rbd-nbd 2026-03-09T19:55:12.321 INFO:teuthology.orchestra.run.vm11.stderr:No packages marked for removal. 2026-03-09T19:55:12.323 INFO:teuthology.orchestra.run.vm11.stdout:Dependencies resolved. 2026-03-09T19:55:12.323 INFO:teuthology.orchestra.run.vm11.stdout:Nothing to do. 2026-03-09T19:55:12.324 INFO:teuthology.orchestra.run.vm11.stdout:Complete! 2026-03-09T19:55:12.326 DEBUG:teuthology.orchestra.run.vm10:> sudo yum clean all 2026-03-09T19:55:12.331 INFO:teuthology.orchestra.run.vm02.stdout:Cache was expired 2026-03-09T19:55:12.331 INFO:teuthology.orchestra.run.vm02.stdout:0 files removed 2026-03-09T19:55:12.344 DEBUG:teuthology.orchestra.run.vm11:> sudo yum clean all 2026-03-09T19:55:12.348 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.463 INFO:teuthology.orchestra.run.vm10.stdout:56 files removed 2026-03-09T19:55:12.476 INFO:teuthology.orchestra.run.vm11.stdout:56 files removed 2026-03-09T19:55:12.490 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.504 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.518 DEBUG:teuthology.orchestra.run.vm10:> sudo yum clean expire-cache 2026-03-09T19:55:12.531 DEBUG:teuthology.orchestra.run.vm11:> sudo yum clean expire-cache 2026-03-09T19:55:12.677 INFO:teuthology.orchestra.run.vm10.stdout:Cache was expired 2026-03-09T19:55:12.677 INFO:teuthology.orchestra.run.vm10.stdout:0 files removed 2026-03-09T19:55:12.697 INFO:teuthology.orchestra.run.vm11.stdout:Cache was expired 2026-03-09T19:55:12.697 INFO:teuthology.orchestra.run.vm11.stdout:0 files removed 2026-03-09T19:55:12.699 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.720 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.720 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm02.local 2026-03-09T19:55:12.721 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm10.local 2026-03-09T19:55:12.721 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm11.local 2026-03-09T19:55:12.721 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.721 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.721 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f /etc/yum.repos.d/ceph.repo 2026-03-09T19:55:12.745 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:55:12.748 DEBUG:teuthology.orchestra.run.vm11:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:55:12.749 DEBUG:teuthology.orchestra.run.vm10:> sudo mv -f /etc/yum/pluginconf.d/priorities.conf.orig /etc/yum/pluginconf.d/priorities.conf 2026-03-09T19:55:12.810 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.812 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.815 DEBUG:teuthology.parallel:result is None 2026-03-09T19:55:12.815 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-09T19:55:12.817 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-09T19:55:12.817 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:55:12.851 DEBUG:teuthology.orchestra.run.vm10:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:55:12.856 DEBUG:teuthology.orchestra.run.vm11:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-09T19:55:12.865 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:55:12.870 INFO:teuthology.orchestra.run.vm10.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:55:12.871 INFO:teuthology.orchestra.run.vm11.stderr:bash: line 1: ntpq: command not found 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:=============================================================================== 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:^+ ntp2.uni-ulm.de 2 6 377 22 +84us[ +169us] +/- 15ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:^+ stratum2-3.NTP.TechFak.N> 2 6 337 21 +1051us[+1136us] +/- 17ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:^* 130.61.89.107 2 6 377 21 -514us[ -428us] +/- 13ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm10.stdout:^+ 141.144.246.224 2 6 377 22 -593us[ -508us] +/- 16ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:^- stratum2-3.NTP.TechFak.N> 2 6 167 18 -87us[ -87us] +/- 15ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:^* 130.61.89.107 2 6 377 20 -276us[ -370us] +/- 13ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:^+ 141.144.246.224 2 6 377 22 -6256us[-6350us] +/- 23ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm02.stdout:^+ ntp2.uni-ulm.de 2 6 377 22 +700us[ +606us] +/- 15ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:=============================================================================== 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:^- stratum2-3.NTP.TechFak.N> 2 6 37 14 +1081us[+1081us] +/- 17ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:^* 130.61.89.107 2 6 377 20 -614us[ -592us] +/- 13ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:^+ 141.144.246.224 2 6 377 88 -352us[ -353us] +/- 16ms 2026-03-09T19:55:13.151 INFO:teuthology.orchestra.run.vm11.stdout:^+ ntp2.uni-ulm.de 2 6 377 23 +593us[ +615us] +/- 15ms 2026-03-09T19:55:13.152 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-09T19:55:13.154 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-09T19:55:13.154 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-09T19:55:13.156 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-09T19:55:13.158 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-09T19:55:13.160 INFO:teuthology.task.internal:Duration was 1151.462834 seconds 2026-03-09T19:55:13.160 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-09T19:55:13.162 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-09T19:55:13.162 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T19:55:13.193 DEBUG:teuthology.orchestra.run.vm10:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T19:55:13.194 DEBUG:teuthology.orchestra.run.vm11:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-09T19:55:13.229 INFO:teuthology.orchestra.run.vm10.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:55:13.233 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:55:13.237 INFO:teuthology.orchestra.run.vm11.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-09T19:55:13.738 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-09T19:55:13.738 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-09T19:55:13.738 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T19:55:13.764 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm10.local 2026-03-09T19:55:13.764 DEBUG:teuthology.orchestra.run.vm10:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T19:55:13.805 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm11.local 2026-03-09T19:55:13.805 DEBUG:teuthology.orchestra.run.vm11:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-09T19:55:13.830 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-09T19:55:13.830 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:13.832 DEBUG:teuthology.orchestra.run.vm10:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:13.847 DEBUG:teuthology.orchestra.run.vm11:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:14.873 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-09T19:55:14.873 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:55:14.875 DEBUG:teuthology.orchestra.run.vm10:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:55:14.877 DEBUG:teuthology.orchestra.run.vm11:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-09T19:55:14.900 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:55:14.901 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:55:14.901 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:55:14.901 INFO:teuthology.orchestra.run.vm11.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm11.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm11.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose/home/ubuntu/cephtest/archive/syslog/kern.log: -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm02.stderr: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T19:55:14.902 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T19:55:14.903 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-09T19:55:14.903 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-09T19:55:14.904 INFO:teuthology.orchestra.run.vm10.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-09T19:55:14.904 INFO:teuthology.orchestra.run.vm10.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-09T19:55:14.904 INFO:teuthology.orchestra.run.vm10.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-09T19:55:15.033 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T19:55:15.034 INFO:teuthology.orchestra.run.vm10.stderr: 98.3% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T19:55:15.253 INFO:teuthology.orchestra.run.vm11.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.5% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-09T19:55:15.255 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-09T19:55:15.257 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-09T19:55:15.257 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T19:55:15.286 DEBUG:teuthology.orchestra.run.vm10:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T19:55:15.309 DEBUG:teuthology.orchestra.run.vm11:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-09T19:55:15.337 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-09T19:55:15.341 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.343 DEBUG:teuthology.orchestra.run.vm10:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.351 DEBUG:teuthology.orchestra.run.vm11:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.368 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-09T19:55:15.376 INFO:teuthology.orchestra.run.vm10.stdout:kernel.core_pattern = core 2026-03-09T19:55:15.407 INFO:teuthology.orchestra.run.vm11.stdout:kernel.core_pattern = core 2026-03-09T19:55:15.421 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.443 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:55:15.443 DEBUG:teuthology.orchestra.run.vm10:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.459 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:55:15.460 DEBUG:teuthology.orchestra.run.vm11:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-09T19:55:15.476 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-09T19:55:15.476 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-09T19:55:15.478 INFO:teuthology.task.internal:Transferring archived files... 2026-03-09T19:55:15.478 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm02 2026-03-09T19:55:15.479 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T19:55:15.511 DEBUG:teuthology.misc:Transferring archived files from vm10:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm10 2026-03-09T19:55:15.511 DEBUG:teuthology.orchestra.run.vm10:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T19:55:15.543 DEBUG:teuthology.misc:Transferring archived files from vm11:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-09_11:23:05-orch-squid-none-default-vps/625/remote/vm11 2026-03-09T19:55:15.543 DEBUG:teuthology.orchestra.run.vm11:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-09T19:55:15.575 INFO:teuthology.task.internal:Removing archive directory... 2026-03-09T19:55:15.575 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T19:55:15.577 DEBUG:teuthology.orchestra.run.vm10:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T19:55:15.585 DEBUG:teuthology.orchestra.run.vm11:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-09T19:55:15.630 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-09T19:55:15.632 INFO:teuthology.task.internal:Not uploading archives. 2026-03-09T19:55:15.632 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-09T19:55:15.635 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-09T19:55:15.635 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T19:55:15.636 DEBUG:teuthology.orchestra.run.vm10:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T19:55:15.644 DEBUG:teuthology.orchestra.run.vm11:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-09T19:55:15.652 INFO:teuthology.orchestra.run.vm02.stdout: 8532141 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 19:55 /home/ubuntu/cephtest 2026-03-09T19:55:15.662 INFO:teuthology.orchestra.run.vm10.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 19:55 /home/ubuntu/cephtest 2026-03-09T19:55:15.687 INFO:teuthology.orchestra.run.vm11.stdout: 8532145 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 9 19:55 /home/ubuntu/cephtest 2026-03-09T19:55:15.688 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-09T19:55:15.693 INFO:teuthology.run:Summary data: description: orch/cephadm/workunits/{0-distro/centos_9.stream agent/off mon_election/classic task/test_host_drain} duration: 1151.4628343582153 flavor: default owner: kyr success: true 2026-03-09T19:55:15.693 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-09T19:55:15.714 INFO:teuthology.run:pass